-
Notifications
You must be signed in to change notification settings - Fork 78
Prototype based on blackhole_fdw and c -> java calls #534
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: master
Are you sure you want to change the base?
Conversation
This is prototype based on the blackhole_fdw. That is - the required method are the minimum required for the FDW to be loaded. There's quite a bit of extra code required if before you can use a real backend that's just happens to also return nothing. I have already implemented that elsewhere but for now it's fine to skip it. That said most of the provided methods call the appropriate java method. THESE ARE NOT STATIC METHODS - it is assumed that everything except for the FDWValidator is associated with a Java object - FDWForeignDataWrapper, FDWServer, FDWTable, or the various internal states. The `fwd_private` fields contain a `jobject` for the java `instance`. It is passed through to java via the JNI call in the same way that we can use `method.invoke(Object obj,...)` in Java. There is also a dummy implementation of the java side of the FDW. It does nothing but log an acknowledgement that the method was called. (Is there a way to call trigger `elog(NOTIFY,...)` ? BONUS - and probably a separate pull request The pljava-so directory contains a Dockerfile and docker-compose.yml file that can be used to create a test image. The makefile(?) needs to add a stanza that calls `docker build ...` - I can provide the details later. For this project it may make more sense to move this to the packaging module - for the other project I'm working on it's entirely standalone so I can do everything in that module. Once there's a docker image I can add an 'example' that uses TestContainers to run some actual java-based tests. KNOWN ISSUES There are a lot. The most basic is that I'm not reusing existing bits of JNI. This should be easy to fix. The most important, and the reason I haven't tried to actually build and run the .so, is that most of these objects need to stay in the persistent memory context since they live beyond the lifetime of the query. I know there's already one in the existing backend code but I don't know if I should use it or a per-FDW one. It's also clear that there needs to be an OID associated with the Foreign Data Wrapper, Server, and Foreign Table objects since they're persistence database objects. It should be possible to reuse existing ones, or at least have a clean way to access them. For now I'm using a simple tree - actually a one-to-one-to-one tree - but while the hooks are there it's not actually executing all of the steps yet. However there is enough that we can create a standalone object for testing. However this has raised a second issue - the method signatures to create a new Server or Foreign Table are identical but both Foreign Data Wrappers and Servers can support multiple children. I'm sure there's an easy way to get the jclass from a jobject but I wanted to focus on other things since we can always fall back to assuming a single matching class. There also needs to be correct mapping between int and jint, char * and jstring, etc. Once we have a solid foundation we can start adding conversions for the rest of the functions/methods. Finally the user OID is available to us but I don't know how to retrieve the desired information from it. For maximum flexibility (and security) we want at least three things: - the database username - the authentication method used - the connection method used (via unix named file, TCP/IP (IP address), etc.) The last item is important in highly secured environments, e.g., some critical operations may be limited to local users - and then the java class may still require additional authentication with something like a Yubico key. (Can you tell my other project is related to encryption keys and signing using harded external devices?) So, overall, this is enough to give a decent representation of what's required to have the C-based FDW call java classes for the actual work. There's quite a bit more work to be done on the FDW side but it's internal bookkeeping and doesn't affect the java SPI. However I don't know what prep work needs to be done beyond what's already done for UDF and UDT, and there's definitely the question of how the different persistent nodes are created and used. I've looked at a few other implementations but this goes a few steps beyond them due to the desire to eventually support multiple servers and tables. FDW Handler?... I've been a bit confused about a single handler when there has always been a desire to provide multiple servers and foreign tables. I think I have an answer though - the two functions are persistent and will have nodes and OIDs associated with them. The initial definition can point to a 'blackhole' Handler but it the handler could be replaced at the Server and Foreign table level. That's the only thing that makes sense since different tables will require different implementations of the planner, scanner, modification methods, etc. It's likely that the Validator can also change since the meaningful values for a Foreign Table may depend on the the specific Server used. BONUS ITEM tada#2 This is WAAAY out there but while digging through the deeply nested structures I came across a function pointer for performance analysis. I don't think it's limited to just FDW - it was pretty deep in the standard structures at this point. On one hand I shudder to think of the impact of proving a JNI binding to allow a java class to collect performance metrics. On the other hand.... That said... I've looked at similar situations in the past and have largely concluded that the best solution - on a single-node system - is to use a named pipe, or IPC if you're more comfortable with it. As long as you're willing to accept data loss if the system is overwhelmed there's not much risk to slamming information into either pseudodevice and relying on a consumer to move the information somewhere else. For instance something like ApacheMQ so that the analysis can be performed on one or more remote machines. However it still got me wondering... and I'm sure there's something similar for row and column-level authorization, auditing, perhaps even encryption.
Here's a bit of info from my other project. First, a quick descent through the obvious tables.
The owner is I'm a bit disappointed that there fdwhandler is only associated with the Foreign Data Wrapper but it's not a problem. The key functions have a More importantly we can create a permanent cache from the table's This would look something like create table jsql.fdw_wrapper_jni (
oid oid not null primary key references pg_catalog.pg_foreign_data_wrapper.oid,
classname text not null
);
create table jsql.fdw_server_jni (
oid oid not null primary key references pg_catalog.pg_foreign_server.oid,
classname text not null
);
create table jsql.fdw_table_jni (
oid oid not null primary key references pg_catalog.pg_foreign_table.oid,
classname text not null,
plan_state_classname text not null,
scan_state_classname text not null
); A quick peek at the functions (fdwhandler, fdwvalidator) are
and the types are
|
After a bit of thought I realized that classnames can stored as standard options - no need for an extra table. I only thought of that because of my initial hope that we could cache jclass and jmethod. It's also occurred to me that I now know how to handle the one-to-many aspect. Java side:
backend site:
I know the java can update the local one-to-many relations when a new object is created but I don't know if there's a clean way to know that one has been deleted. I think most of the FDWRoutine methods already use the I'll try to submit a revision later today. |
Quick update since I wasn't at a pull-request-ready stage but hope to have something in the next few days. I've refactored the java side extensively to match my new information about what's created when. That said I still have some questions that can only be resolved by running some actual code. So the current plan is two-pronged (and hence the delay.) The first is incorporate big chunks of my existing project that does successfully load a FDW. That FDW doesn't do much - but it doesn't blow chunks when I execute The second is to have a preliminary implementation of more of the JNI methods while using Java constructs like Finally I'll add huge #ifdef NEVER
// any JNI code that I'm not 100% confident on
#endif and verify that I can build a working FDW. That will make further development much easier since we'll have immediate feedback whether the changes 1) broke anything and 2) worked as expected. I mentioned earlier that there's a standard location where you can put initialization scripts (both shell and sql) so the docker image can immediately create the extension, fdw, foreign user, server, and table. Or define a stored procedure to do this so we can see any NOTIFY messages. Java implementation snippetI have a tentative implementation of a not-quite-blackhole table. It should give you a feel for what I'm thinking of, but of course many details are still up in the air. public class BlackholeTable {
private final BlackholeServer server;'
private final Map<String, String> options;
private final List<Map<String, Object>> data = new ArrayList<>();
public BlackholeTable(BlackholeServer server, Map<String, String> options) {
this.server = server;
this.options = new LinkedHashMap<>(options);
// do something based on options
// e.g., they may define a mapping from the database column name with a class's field name and/or accessor.
..
// initialize data with predefined values.
}
// Note: this should also provide current database user and foreign user information
// for any additional authorization checks.
public BlackholeTableScanState newScanState(this, data) {
return new BlackholeTableScanState(this, data);
}
}
public class BlackholeTableScanState {
private final List<Map<String, Object>> data;
private Iterator<Map<String, Object> iter;
private isOpen;
public BlackholeTableScanState(BlackholeTable table, List<Map<String, Object>> data) {
this.table = table;
this.data = new Ar rayList<>(data);
this.isOpen = false;
}
public boolean open() {
if (isOpen) { } // log this for now
if (iter != null) {
// log this for now
iter = null;
}
// do some safe stuff
if (explainOnly) {
return false;
}
// if necessary open and close external resources
// ...
isOpen = true;
return true;
}
public Map<String, Object> next() {
if (!isOpen) { } // log this for now
if (iter == null) {
iter = data.iterator();
}
if (iter.hasNext()) {
return new LinkedHashMap<>(iter.next());
} else {
return Collections.emptyMap();
}
}
public void reset() {
if (!isOpen) { } // log this for now
iter = data.iterator();
}
public voiid close() {
if (!isOpen) { } // log this for now
iter = null
data.clear();
}
public void explain() {
// provide information about the table, server, wrapper, possibly user
}
} Java annotationsThe prior code assumes a very simple relationship, e.g., the map's keys have the same name as the foreign table's columns. This puts a nontrivial load on the developer. This could be hidden with annotations. The class-level annotation would provide the table options (e.g., URL) while the field- or method-level annotations could provide an additional table option for mapping column names. This would make implementations cleaner (more DB-agnostic) and improve the potential for reuse. Further abstractionWe could make the implementation of the There may be a solution since Stream<T> generate(Supplier<T> s); and we would be covered if the This seems like weird requirement but it would allow the framework to hide massive amounts of implementation details from the java developer. |
I'll look forward to seeing that. I had questions about the way it is organized here. As a kind of combination of useful information and cautionary tale, I offer this youtube link. It's a 25-minute talk that was given at Postgres Extensions Day in Montréal earlier this month. The presenter describes a project where he used C, Java, and JNI inside a PostgreSQL backend, and he goes into everything he had to invent solutions for in dealing with JNI, memory management, signals, Java multithreading in single-threaded PostgreSQL, and so on. He was one of three developers on the project, which they started a year ago, and somehow three developers worked for a year on a Java-with-JNI-in-PostgreSQL project and prepared a 25-minute talk about it without ever having looked at how PL/Java addresses any of those things, and so it was a 25-minute talk given in the year 2025 that could have come straight from the archived original documentation in the PL/Java repo from the year 2005. I'm going to seriously recommend that you perhaps look at both: his video, and the archived PL/Java docs, in particular After time-traveling back to the archived docs, you will probably want to get a better sense of what in PL/Java's internals still fairly closely resembles what Thomas wrote 20 years ago, and which of his solutions have more significantly evolved. For starters, I would recommend looking at two specific things: While looking at These are not things you can hack on PL/Java (in the C parts, anyway) without knowing. |
This is prototype based on the blackhole_fdw. That is - the required method are the minimum required for the FDW to be loaded. There's quite a bit of extra code required if before you can use a real backend that's just happens to also return nothing. I have already implemented that elsewhere but for now it's fine to skip it.
That said most of the provided methods call the appropriate java method. THESE ARE NOT STATIC METHODS - it is assumed that everything except for the FDWValidator is associated with a Java object - FDWForeignDataWrapper, FDWServer, FDWTable, or the various internal states.
The
fwd_private
fields contain ajobject
for the javainstance
. It is passed through to java via the JNI call in the same way that we can usemethod.invoke(Object obj,...)
in Java.There is also a dummy implementation of the java side of the FDW. It does nothing but log an acknowledgement that the method was called. (Is there a way to call trigger
elog(NOTIFY,...)
?BONUS - and probably a separate pull request
The pljava-so directory contains a Dockerfile and docker-compose.yml file that can be used to create a test image. The makefile(?) needs to add a stanza that calls
docker build ...
- I can provide the details later. For this project it may make more sense to move this to the packaging module - for the other project I'm working on it's entirely standalone so I can do everything in that module.Once there's a docker image I can add an 'example' that uses TestContainers to run some actual java-based tests.
KNOWN ISSUES
There are a lot.
The most basic is that I'm not reusing existing bits of JNI. This should be easy to fix.
The most important, and the reason I haven't tried to actually build and run the .so, is that most of these objects need to stay in the persistent memory context since they live beyond the lifetime of the query. I know there's already one in the existing backend code but I don't know if I should use it or a per-FDW one.
It's also clear that there needs to be an OID associated with the Foreign Data Wrapper, Server, and Foreign Table objects since they're persistence database objects. It should be possible to reuse existing ones, or at least have a clean way to access them.
For now I'm using a simple tree - actually a one-to-one-to-one tree - but while the hooks are there it's not actually executing all of the steps yet. However there is enough that we can create a standalone object for testing.
However this has raised a second issue - the method signatures to create a new Server or Foreign Table are identical but both Foreign Data Wrappers and Servers can support multiple children. I'm sure there's an easy way to get the jclass from a jobject but I wanted to focus on other things since we can always fall back to assuming a single matching class.
There also needs to be correct mapping between int and jint, char * and jstring, etc. Once we have a solid foundation we can start adding conversions for the rest of the functions/methods.
Finally the user OID is available to us but I don't know how to retrieve the desired information from it. For maximum flexibility (and security) we want at least three things:
The last item is important in highly secured environments, e.g., some critical operations may be limited to local users - and then the java class may still require additional authentication with something like a Yubico key. (Can you tell my other project is related to encryption keys and signing using harded external devices?)
So, overall, this is enough to give a decent representation of what's required to have the C-based FDW call java classes for the actual work. There's quite a bit more work to be done on the FDW side but it's internal bookkeeping and doesn't affect the java SPI.
However I don't know what prep work needs to be done beyond what's already done for UDF and UDT, and there's definitely the question of how the different persistent nodes are created and used. I've looked at a few other implementations but this goes a few steps beyond them due to the desire to eventually support multiple servers and tables.
FDW Handler?...
I've been a bit confused about a single handler when there has always been a desire to provide multiple servers and foreign tables. I think I have an answer though - the two functions are persistent and will have nodes and OIDs associated with them. The initial definition can point to a 'blackhole' Handler but it the handler could be replaced at the Server and Foreign table level. That's the only thing that makes sense since different tables will require different implementations of the planner, scanner, modification methods, etc.
It's likely that the Validator can also change since the meaningful values for a Foreign Table may depend on the the specific Server used.
BONUS ITEM #2
This is WAAAY out there but while digging through the deeply nested structures I came across a function pointer for performance analysis. I don't think it's limited to just FDW - it was pretty deep in the standard structures at this point.
On one hand I shudder to think of the impact of proving a JNI binding to allow a java class to collect performance metrics. On the other hand....
That said... I've looked at similar situations in the past and have largely concluded that the best solution - on a single-node system - is to use a named pipe, or IPC if you're more comfortable with it. As long as you're willing to accept data loss if the system is overwhelmed there's not much risk to slamming information into either pseudodevice and relying on a consumer to move the information somewhere else. For instance something like ApacheMQ so that the analysis can be performed on one or more remote machines.
However it still got me wondering... and I'm sure there's something similar for row and column-level authorization, auditing, perhaps even encryption.