-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat(SDK) Add java SDK for structuredProperty entity PATCH + CRUD examples #10823
feat(SDK) Add java SDK for structuredProperty entity PATCH + CRUD examples #10823
Conversation
WalkthroughThis update improves the handling of structured properties in the DataHub project. It introduces new methods for setting and removing string and number properties in structured properties. It also expands the testing infrastructure and provides examples demonstrating how to update and upsert structured properties for various entities. Changes
Poem
Tip AI model upgrade
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 4
Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Files selected for processing (5)
- entity-registry/src/main/java/com/linkedin/metadata/aspect/patch/builder/StructuredPropertiesPatchBuilder.java (3 hunks)
- entity-registry/src/main/java/com/linkedin/metadata/aspect/patch/builder/StructuredPropertyDefinitionPatchBuilder.java (1 hunks)
- metadata-integration/java/datahub-client/src/test/java/datahub/client/patch/PatchTest.java (3 hunks)
- metadata-integration/java/examples/src/main/java/io/datahubproject/examples/DatasetStructuredPropertiesUpdate.java (1 hunks)
- metadata-integration/java/examples/src/main/java/io/datahubproject/examples/StructuredPropertyUpsert.java (1 hunks)
Additional comments not posted (14)
entity-registry/src/main/java/com/linkedin/metadata/aspect/patch/builder/StructuredPropertiesPatchBuilder.java (4)
Line range hint
19-34
: LGTM!The method correctly removes a structured property by its URN.
53-69
: LGTM!The method correctly sets a number property with a list of integer values.
90-106
: LGTM!The method correctly sets a string property with a list of string values.
Line range hint
137-145
: LGTM!The methods correctly return the aspect name and entity type.
entity-registry/src/main/java/com/linkedin/metadata/aspect/patch/builder/StructuredPropertyDefinitionPatchBuilder.java (6)
34-41
: LGTM!The method correctly sets the qualified name for a structured property.
43-50
: LGTM!The method correctly sets the display name for a structured property.
52-59
: LGTM!The method correctly sets the value type for a structured property.
61-75
: LGTM!The method correctly sets the type qualifier for a structured property.
77-92
: LGTM!The method correctly adds an allowed value for a structured property.
94-102
: LGTM!The method correctly sets the cardinality for a structured property.
metadata-integration/java/datahub-client/src/test/java/datahub/client/patch/PatchTest.java (4)
652-654
: Verify the reason for ignoring the test.The test method is marked with
@Ignore
. Ensure that this is intentional and not an oversight.
696-696
: Ensure the test should be active.The test method is not marked with
@Ignore
. Ensure that this is intentional and not an oversight.
697-718
: LGTM! Ensure the correctness of the patch builder logic.The logic for building the patch looks good. Verify that the
StructuredPropertiesPatchBuilder
is functioning as expected.
655-693
: LGTM! Ensure the correctness of the patch builder logic.The logic for building the patch looks good. Verify that the
StructuredPropertyDefinitionPatchBuilder
is functioning as expected.
public static void main(String[] args) | ||
throws IOException, ExecutionException, InterruptedException { | ||
|
||
// Adding a structured property with a single string value | ||
MetadataChangeProposal mcp1 = | ||
new StructuredPropertiesPatchBuilder() | ||
.urn( | ||
UrnUtils.getUrn( | ||
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)")) | ||
.setStringProperty( | ||
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.privacy.retentionTime"), "30") | ||
.build(); | ||
|
||
String token = ""; | ||
RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token)); | ||
Future<MetadataWriteResponse> response1 = emitter.emit(mcp1, null); | ||
System.out.println(response1.get().getResponseContent()); | ||
|
||
// Adding a structured property with a list of string values | ||
List<String> values = new ArrayList<>(); | ||
values.add("30"); | ||
values.add("90"); | ||
MetadataChangeProposal mcp2 = | ||
new StructuredPropertiesPatchBuilder() | ||
.urn( | ||
UrnUtils.getUrn( | ||
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)")) | ||
.setStringProperty( | ||
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.privacy.retentionTime"), values) | ||
.build(); | ||
|
||
Future<MetadataWriteResponse> response2 = emitter.emit(mcp2, null); | ||
System.out.println(response2.get().getResponseContent()); | ||
|
||
// Adding a structured property with a single number value | ||
MetadataChangeProposal mcp3 = | ||
new StructuredPropertiesPatchBuilder() | ||
.urn( | ||
UrnUtils.getUrn( | ||
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)")) | ||
.setNumberProperty( | ||
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.dataManagement.replicationSLA"), | ||
3456) | ||
.build(); | ||
|
||
Future<MetadataWriteResponse> response3 = emitter.emit(mcp3, null); | ||
System.out.println(response3.get().getResponseContent()); | ||
|
||
// Removing a structured property from a dataset | ||
MetadataChangeProposal mcp4 = | ||
new StructuredPropertiesPatchBuilder() | ||
.urn( | ||
UrnUtils.getUrn( | ||
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)")) | ||
.removeProperty( | ||
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.dataManagement.replicationSLA")) | ||
.build(); | ||
|
||
Future<MetadataWriteResponse> response4 = emitter.emit(mcp4, null); | ||
System.out.println(response4.get().getResponseContent()); | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Add error handling and logging.
The main
method lacks error handling and logging. Adding these will enhance maintainability and debugging.
public static void main(String[] args)
throws IOException, ExecutionException, InterruptedException {
try {
// Adding a structured property with a single string value
MetadataChangeProposal mcp1 =
new StructuredPropertiesPatchBuilder()
.urn(
UrnUtils.getUrn(
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)"))
.setStringProperty(
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.privacy.retentionTime"), "30")
.build();
String token = "";
RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token));
Future<MetadataWriteResponse> response1 = emitter.emit(mcp1, null);
System.out.println(response1.get().getResponseContent());
// Adding a structured property with a list of string values
List<String> values = new ArrayList<>();
values.add("30");
values.add("90");
MetadataChangeProposal mcp2 =
new StructuredPropertiesPatchBuilder()
.urn(
UrnUtils.getUrn(
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)"))
.setStringProperty(
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.privacy.retentionTime"), values)
.build();
Future<MetadataWriteResponse> response2 = emitter.emit(mcp2, null);
System.out.println(response2.get().getResponseContent());
// Adding a structured property with a single number value
MetadataChangeProposal mcp3 =
new StructuredPropertiesPatchBuilder()
.urn(
UrnUtils.getUrn(
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)"))
.setNumberProperty(
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.dataManagement.replicationSLA"),
3456)
.build();
Future<MetadataWriteResponse> response3 = emitter.emit(mcp3, null);
System.out.println(response3.get().getResponseContent());
// Removing a structured property from a dataset
MetadataChangeProposal mcp4 =
new StructuredPropertiesPatchBuilder()
.urn(
UrnUtils.getUrn(
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)"))
.removeProperty(
UrnUtils.getUrn("urn:li:structuredProperty:io.acryl.dataManagement.replicationSLA"))
.build();
Future<MetadataWriteResponse> response4 = emitter.emit(mcp4, null);
System.out.println(response4.get().getResponseContent());
} catch (Exception e) {
System.err.println("An error occurred: " + e.getMessage());
e.printStackTrace();
}
}
public StructuredPropertiesPatchBuilder setNumberProperty( | ||
@Nonnull Urn propertyUrn, @Nullable Integer propertyValue) { | ||
ValueNode propertyValueNode = instance.numberNode((Integer) propertyValue); | ||
ObjectNode value = instance.objectNode(); | ||
value.put(URN_KEY, propertyUrn.toString()); | ||
ObjectNode newProperty = instance.objectNode(); | ||
newProperty.put(URN_KEY, propertyUrn.toString()); | ||
|
||
ArrayNode valuesNode = instance.arrayNode(); | ||
ObjectNode propertyValueNode = instance.objectNode(); | ||
propertyValueNode.set("double", instance.numberNode(propertyValue)); | ||
valuesNode.add(propertyValueNode); | ||
newProperty.set(VALUES_KEY, valuesNode); | ||
|
||
pathValues.add( | ||
ImmutableTriple.of( | ||
PatchOperationType.ADD.getValue(), BASE_PATH + "/" + propertyUrn, propertyValueNode)); | ||
PatchOperationType.ADD.getValue(), BASE_PATH + "/" + propertyUrn, newProperty)); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Add null check for propertyValue
.
The method can be improved by adding a null check for propertyValue
to handle potential null values gracefully.
public StructuredPropertiesPatchBuilder setNumberProperty(
@Nonnull Urn propertyUrn, @Nullable Integer propertyValue) {
ObjectNode newProperty = instance.objectNode();
newProperty.put(URN_KEY, propertyUrn.toString());
ArrayNode valuesNode = instance.arrayNode();
if (propertyValue != null) {
ObjectNode propertyValueNode = instance.objectNode();
propertyValueNode.set("double", instance.numberNode(propertyValue));
valuesNode.add(propertyValueNode);
}
newProperty.set(VALUES_KEY, valuesNode);
pathValues.add(
ImmutableTriple.of(
PatchOperationType.ADD.getValue(), BASE_PATH + "/" + propertyUrn, newProperty));
return this;
}
public StructuredPropertiesPatchBuilder setStringProperty( | ||
@Nonnull Urn propertyUrn, @Nullable String propertyValue) { | ||
ObjectNode newProperty = instance.objectNode(); | ||
newProperty.put(URN_KEY, propertyUrn.toString()); | ||
|
||
ArrayNode valuesNode = instance.arrayNode(); | ||
ObjectNode propertyValueNode = instance.objectNode(); | ||
propertyValueNode.set("string", instance.textNode(propertyValue)); | ||
valuesNode.add(propertyValueNode); | ||
newProperty.set(VALUES_KEY, valuesNode); | ||
|
||
pathValues.add( | ||
ImmutableTriple.of( | ||
PatchOperationType.ADD.getValue(), | ||
BASE_PATH + "/" + propertyUrn + "/" + String.valueOf(propertyValue), | ||
propertyValueNode)); | ||
PatchOperationType.ADD.getValue(), BASE_PATH + "/" + propertyUrn, newProperty)); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Add null check for propertyValue
.
The method can be improved by adding a null check for propertyValue
to handle potential null values gracefully.
public StructuredPropertiesPatchBuilder setStringProperty(
@Nonnull Urn propertyUrn, @Nullable String propertyValue) {
ObjectNode newProperty = instance.objectNode();
newProperty.put(URN_KEY, propertyUrn.toString());
ArrayNode valuesNode = instance.arrayNode();
if (propertyValue != null) {
ObjectNode propertyValueNode = instance.objectNode();
propertyValueNode.set("string", instance.textNode(propertyValue));
valuesNode.add(propertyValueNode);
}
newProperty.set(VALUES_KEY, valuesNode);
pathValues.add(
ImmutableTriple.of(
PatchOperationType.ADD.getValue(), BASE_PATH + "/" + propertyUrn, newProperty));
return this;
}
public static void main(String[] args) | ||
throws IOException, ExecutionException, InterruptedException { | ||
// open ended string structured property on datasets and dataFlows | ||
MetadataChangeProposal mcp1 = | ||
new StructuredPropertyDefinitionPatchBuilder() | ||
.urn( | ||
UrnUtils.getUrn( | ||
"urn:li:structuredProperty:testString")) // use existing urn for update, new urn | ||
// for new property | ||
.setQualifiedName("io.acryl.testString") | ||
.setDisplayName("Open Ended String") | ||
.setValueType("urn:li:dataType:datahub.string") | ||
.setCardinality(PropertyCardinality.SINGLE) | ||
.addEntityType("urn:li:entityType:datahub.dataset") | ||
.addEntityType("urn:li:entityType:datahub.dataFlow") | ||
.setDescription("test description for open ended string") | ||
.setImmutable(true) | ||
.build(); | ||
|
||
String token = ""; | ||
RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token)); | ||
Future<MetadataWriteResponse> response1 = emitter.emit(mcp1, null); | ||
System.out.println(response1.get().getResponseContent()); | ||
|
||
// Next, let's make a property that allows for multiple datahub entity urns as values | ||
// This example property could be used to reference other users or groups in datahub | ||
StringArrayMap typeQualifier = new StringArrayMap(); | ||
typeQualifier.put( | ||
"allowedTypes", | ||
new StringArray( | ||
"urn:li:entityType:datahub.corpuser", "urn:li:entityType:datahub.corpGroup")); | ||
|
||
MetadataChangeProposal mcp2 = | ||
new StructuredPropertyDefinitionPatchBuilder() | ||
.urn(UrnUtils.getUrn("urn:li:structuredProperty:dataSteward")) | ||
.setQualifiedName("io.acryl.dataManagement.dataSteward") | ||
.setDisplayName("Data Steward") | ||
.setValueType("urn:li:dataType:datahub.urn") | ||
.setTypeQualifier(typeQualifier) | ||
.setCardinality(PropertyCardinality.MULTIPLE) | ||
.addEntityType("urn:li:entityType:datahub.dataset") | ||
.setDescription( | ||
"The data stewards of this asset are in charge of ensuring data cleanliness and governance") | ||
.setImmutable(true) | ||
.build(); | ||
|
||
Future<MetadataWriteResponse> response2 = emitter.emit(mcp2, null); | ||
System.out.println(response2.get().getResponseContent()); | ||
|
||
// Finally, let's make a single select number property with a few allowed options | ||
PropertyValue propertyValue1 = new PropertyValue(); | ||
PrimitivePropertyValue value1 = new PrimitivePropertyValue(); | ||
value1.setDouble(30.0); | ||
propertyValue1.setDescription( | ||
"30 days, usually reserved for datasets that are ephemeral and contain pii"); | ||
propertyValue1.setValue(value1); | ||
PropertyValue propertyValue2 = new PropertyValue(); | ||
PrimitivePropertyValue value2 = new PrimitivePropertyValue(); | ||
value2.setDouble(90.0); | ||
propertyValue2.setDescription( | ||
"Use this for datasets that drive monthly reporting but contain pii"); | ||
propertyValue2.setValue(value2); | ||
|
||
MetadataChangeProposal mcp3 = | ||
new StructuredPropertyDefinitionPatchBuilder() | ||
.urn(UrnUtils.getUrn("urn:li:structuredProperty:replicationSLA")) | ||
.setQualifiedName("io.acryl.dataManagement.replicationSLA") | ||
.setDisplayName("Replication SLA") | ||
.setValueType("urn:li:dataType:datahub.string") | ||
.addAllowedValue(propertyValue1) | ||
.addAllowedValue(propertyValue2) | ||
.setCardinality(PropertyCardinality.SINGLE) | ||
.addEntityType("urn:li:entityType:datahub.dataset") | ||
.setDescription( | ||
"SLA for how long data can be delayed before replicating to the destination cluster") | ||
.setImmutable(false) | ||
.build(); | ||
|
||
Future<MetadataWriteResponse> response3 = emitter.emit(mcp3, null); | ||
System.out.println(response3.get().getResponseContent()); | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Add error handling and logging.
The main
method lacks error handling and logging. Adding these will enhance maintainability and debugging.
public static void main(String[] args)
throws IOException, ExecutionException, InterruptedException {
try {
// open ended string structured property on datasets and dataFlows
MetadataChangeProposal mcp1 =
new StructuredPropertyDefinitionPatchBuilder()
.urn(
UrnUtils.getUrn(
"urn:li:structuredProperty:testString")) // use existing urn for update, new urn
// for new property
.setQualifiedName("io.acryl.testString")
.setDisplayName("Open Ended String")
.setValueType("urn:li:dataType:datahub.string")
.setCardinality(PropertyCardinality.SINGLE)
.addEntityType("urn:li:entityType:datahub.dataset")
.addEntityType("urn:li:entityType:datahub.dataFlow")
.setDescription("test description for open ended string")
.setImmutable(true)
.build();
String token = "";
RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token));
Future<MetadataWriteResponse> response1 = emitter.emit(mcp1, null);
System.out.println(response1.get().getResponseContent());
// Next, let's make a property that allows for multiple datahub entity urns as values
// This example property could be used to reference other users or groups in datahub
StringArrayMap typeQualifier = new StringArrayMap();
typeQualifier.put(
"allowedTypes",
new StringArray(
"urn:li:entityType:datahub.corpuser", "urn:li:entityType:datahub.corpGroup"));
MetadataChangeProposal mcp2 =
new StructuredPropertyDefinitionPatchBuilder()
.urn(UrnUtils.getUrn("urn:li:structuredProperty:dataSteward"))
.setQualifiedName("io.acryl.dataManagement.dataSteward")
.setDisplayName("Data Steward")
.setValueType("urn:li:dataType:datahub.urn")
.setTypeQualifier(typeQualifier)
.setCardinality(PropertyCardinality.MULTIPLE)
.addEntityType("urn:li:entityType:datahub.dataset")
.setDescription(
"The data stewards of this asset are in charge of ensuring data cleanliness and governance")
.setImmutable(true)
.build();
Future<MetadataWriteResponse> response2 = emitter.emit(mcp2, null);
System.out.println(response2.get().getResponseContent());
// Finally, let's make a single select number property with a few allowed options
PropertyValue propertyValue1 = new PropertyValue();
PrimitivePropertyValue value1 = new PrimitivePropertyValue();
value1.setDouble(30.0);
propertyValue1.setDescription(
"30 days, usually reserved for datasets that are ephemeral and contain pii");
propertyValue1.setValue(value1);
PropertyValue propertyValue2 = new PropertyValue();
PrimitivePropertyValue value2 = new PrimitivePropertyValue();
value2.setDouble(90.0);
propertyValue2.setDescription(
"Use this for datasets that drive monthly reporting but contain pii");
propertyValue2.setValue(value2);
MetadataChangeProposal mcp3 =
new StructuredPropertyDefinitionPatchBuilder()
.urn(UrnUtils.getUrn("urn:li:structuredProperty:replicationSLA"))
.setQualifiedName("io.acryl.dataManagement.replicationSLA")
.setDisplayName("Replication SLA")
.setValueType("urn:li:dataType:datahub.string")
.addAllowedValue(propertyValue1)
.addAllowedValue(propertyValue2)
.setCardinality(PropertyCardinality.SINGLE)
.addEntityType("urn:li:entityType:datahub.dataset")
.setDescription(
"SLA for how long data can be delayed before replicating to the destination cluster")
.setImmutable(false)
.build();
Future<MetadataWriteResponse> response3 = emitter.emit(mcp3, null);
System.out.println(response3.get().getResponseContent());
} catch (Exception e) {
System.err.println("An error occurred: " + e.getMessage());
e.printStackTrace();
}
}
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 2
Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Files selected for processing (1)
- metadata-integration/java/datahub-client/src/test/java/datahub/client/patch/PatchTest.java (4 hunks)
Additional comments not posted (1)
metadata-integration/java/datahub-client/src/test/java/datahub/client/patch/PatchTest.java (1)
20-21
: Imports are necessary and correctly used.The added imports for
StringArray
,StringArrayMap
,StructuredPropertiesPatchBuilder
,StructuredPropertyDefinitionPatchBuilder
,PrimitivePropertyValue
,PropertyCardinality
, andPropertyValue
are required and correctly used in the newly added tests.Also applies to: 35-36, 40-42
@Test | ||
@Ignore | ||
public void testLocalStructuredPropertiesUpdate() { | ||
try { | ||
MetadataChangeProposal mcp = | ||
new StructuredPropertiesPatchBuilder() | ||
.urn( | ||
UrnUtils.getUrn( | ||
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)")) | ||
.setNumberProperty( | ||
UrnUtils.getUrn( | ||
"urn:li:structuredProperty:io.acryl.dataManagement.replicationSLA"), | ||
3456) | ||
.build(); | ||
|
||
String token = ""; | ||
RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token)); | ||
Future<MetadataWriteResponse> response = emitter.emit(mcp, null); | ||
System.out.println(response.get().getResponseContent()); | ||
|
||
} catch (IOException | ExecutionException | InterruptedException e) { | ||
System.out.println(Arrays.asList(e.getStackTrace())); | ||
} | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ensure proper handling of RestEmitter
resource and token usage.
The RestEmitter
resource should be properly closed after use to avoid potential resource leaks. Additionally, the token usage should be verified for correctness.
try (RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token))) {
...
} catch (IOException | ExecutionException | InterruptedException e) {
System.out.println(Arrays.asList(e.getStackTrace()));
}
Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
@Test | |
@Ignore | |
public void testLocalStructuredPropertiesUpdate() { | |
try { | |
MetadataChangeProposal mcp = | |
new StructuredPropertiesPatchBuilder() | |
.urn( | |
UrnUtils.getUrn( | |
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)")) | |
.setNumberProperty( | |
UrnUtils.getUrn( | |
"urn:li:structuredProperty:io.acryl.dataManagement.replicationSLA"), | |
3456) | |
.build(); | |
String token = ""; | |
RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token)); | |
Future<MetadataWriteResponse> response = emitter.emit(mcp, null); | |
System.out.println(response.get().getResponseContent()); | |
} catch (IOException | ExecutionException | InterruptedException e) { | |
System.out.println(Arrays.asList(e.getStackTrace())); | |
} | |
} | |
@Test | |
@Ignore | |
public void testLocalStructuredPropertiesUpdate() { | |
try { | |
MetadataChangeProposal mcp = | |
new StructuredPropertiesPatchBuilder() | |
.urn( | |
UrnUtils.getUrn( | |
"urn:li:dataset:(urn:li:dataPlatform:hive,SampleCypressHiveDataset,PROD)")) | |
.setNumberProperty( | |
UrnUtils.getUrn( | |
"urn:li:structuredProperty:io.acryl.dataManagement.replicationSLA"), | |
3456) | |
.build(); | |
String token = ""; | |
try (RestEmitter emitter = RestEmitter.create(b -> b.server("http://localhost:8080").token(token))) { | |
Future<MetadataWriteResponse> response = emitter.emit(mcp, null); | |
System.out.println(response.get().getResponseContent()); | |
} | |
} catch (IOException | ExecutionException | InterruptedException e) { | |
System.out.println(Arrays.asList(e.getStackTrace())); | |
} | |
} |
@Test | ||
@Ignore | ||
public void testLocalStructuredPropertyDefinitionAdd() { | ||
RestEmitter restEmitter = new RestEmitter(RestEmitterConfig.builder().build()); | ||
try { | ||
StringArrayMap typeQualifier = new StringArrayMap(); | ||
typeQualifier.put( | ||
"allowedTypes", | ||
new StringArray( | ||
"urn:li:entityType:datahub.corpuser", "urn:li:entityType:datahub.corpGroup")); | ||
PropertyValue propertyValue1 = new PropertyValue(); | ||
PrimitivePropertyValue value1 = new PrimitivePropertyValue(); | ||
value1.setString("test value 1"); | ||
propertyValue1.setValue(value1); | ||
PropertyValue propertyValue2 = new PropertyValue(); | ||
PrimitivePropertyValue value2 = new PrimitivePropertyValue(); | ||
value2.setString("test value 2"); | ||
propertyValue2.setValue(value2); | ||
|
||
MetadataChangeProposal structuredPropertyDefinitionPatch = | ||
new StructuredPropertyDefinitionPatchBuilder() | ||
.urn(UrnUtils.getUrn("urn:li:structuredProperty:123456")) | ||
.setQualifiedName("test.testing.123") | ||
.setDisplayName("Test Display Name") | ||
.setValueType("urn:li:dataType:datahub.urn") | ||
.setTypeQualifier(typeQualifier) | ||
.addAllowedValue(propertyValue1) | ||
.addAllowedValue(propertyValue2) | ||
.setCardinality(PropertyCardinality.MULTIPLE) | ||
.addEntityType("urn:li:entityType:datahub.dataFlow") | ||
.setDescription("test description") | ||
.setImmutable(true) | ||
.build(); | ||
|
||
Future<MetadataWriteResponse> response = restEmitter.emit(structuredPropertyDefinitionPatch); | ||
|
||
System.out.println(response.get().getResponseContent()); | ||
|
||
} catch (IOException | ExecutionException | InterruptedException e) { | ||
System.out.println(Arrays.asList(e.getStackTrace())); | ||
} | ||
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ensure proper handling of RestEmitter
resource.
The RestEmitter
resource should be properly closed after use to avoid potential resource leaks.
try (RestEmitter restEmitter = new RestEmitter(RestEmitterConfig.builder().build())) {
...
} catch (IOException | ExecutionException | InterruptedException e) {
System.out.println(Arrays.asList(e.getStackTrace()));
}
Committable suggestion
‼️ IMPORTANT
Carefully review the code before committing. Ensure that it accurately replaces the highlighted code, contains no missing lines, and has no issues with indentation. Thoroughly test & benchmark the code to ensure it meets the requirements.
@Test | |
@Ignore | |
public void testLocalStructuredPropertyDefinitionAdd() { | |
RestEmitter restEmitter = new RestEmitter(RestEmitterConfig.builder().build()); | |
try { | |
StringArrayMap typeQualifier = new StringArrayMap(); | |
typeQualifier.put( | |
"allowedTypes", | |
new StringArray( | |
"urn:li:entityType:datahub.corpuser", "urn:li:entityType:datahub.corpGroup")); | |
PropertyValue propertyValue1 = new PropertyValue(); | |
PrimitivePropertyValue value1 = new PrimitivePropertyValue(); | |
value1.setString("test value 1"); | |
propertyValue1.setValue(value1); | |
PropertyValue propertyValue2 = new PropertyValue(); | |
PrimitivePropertyValue value2 = new PrimitivePropertyValue(); | |
value2.setString("test value 2"); | |
propertyValue2.setValue(value2); | |
MetadataChangeProposal structuredPropertyDefinitionPatch = | |
new StructuredPropertyDefinitionPatchBuilder() | |
.urn(UrnUtils.getUrn("urn:li:structuredProperty:123456")) | |
.setQualifiedName("test.testing.123") | |
.setDisplayName("Test Display Name") | |
.setValueType("urn:li:dataType:datahub.urn") | |
.setTypeQualifier(typeQualifier) | |
.addAllowedValue(propertyValue1) | |
.addAllowedValue(propertyValue2) | |
.setCardinality(PropertyCardinality.MULTIPLE) | |
.addEntityType("urn:li:entityType:datahub.dataFlow") | |
.setDescription("test description") | |
.setImmutable(true) | |
.build(); | |
Future<MetadataWriteResponse> response = restEmitter.emit(structuredPropertyDefinitionPatch); | |
System.out.println(response.get().getResponseContent()); | |
} catch (IOException | ExecutionException | InterruptedException e) { | |
System.out.println(Arrays.asList(e.getStackTrace())); | |
} | |
} | |
@Test | |
@Ignore | |
public void testLocalStructuredPropertyDefinitionAdd() { | |
try (RestEmitter restEmitter = new RestEmitter(RestEmitterConfig.builder().build())) { | |
StringArrayMap typeQualifier = new StringArrayMap(); | |
typeQualifier.put( | |
"allowedTypes", | |
new StringArray( | |
"urn:li:entityType:datahub.corpuser", "urn:li:entityType:datahub.corpGroup")); | |
PropertyValue propertyValue1 = new PropertyValue(); | |
PrimitivePropertyValue value1 = new PrimitivePropertyValue(); | |
value1.setString("test value 1"); | |
propertyValue1.setValue(value1); | |
PropertyValue propertyValue2 = new PropertyValue(); | |
PrimitivePropertyValue value2 = new PrimitivePropertyValue(); | |
value2.setString("test value 2"); | |
propertyValue2.setValue(value2); | |
MetadataChangeProposal structuredPropertyDefinitionPatch = | |
new StructuredPropertyDefinitionPatchBuilder() | |
.urn(UrnUtils.getUrn("urn:li:structuredProperty:123456")) | |
.setQualifiedName("test.testing.123") | |
.setDisplayName("Test Display Name") | |
.setValueType("urn:li:dataType:datahub.urn") | |
.setTypeQualifier(typeQualifier) | |
.addAllowedValue(propertyValue1) | |
.addAllowedValue(propertyValue2) | |
.setCardinality(PropertyCardinality.MULTIPLE) | |
.addEntityType("urn:li:entityType:datahub.dataFlow") | |
.setDescription("test description") | |
.setImmutable(true) | |
.build(); | |
Future<MetadataWriteResponse> response = restEmitter.emit(structuredPropertyDefinitionPatch); | |
System.out.println(response.get().getResponseContent()); | |
} catch (IOException | ExecutionException | InterruptedException e) { | |
System.out.println(Arrays.asList(e.getStackTrace())); | |
} | |
} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
Review details
Configuration used: CodeRabbit UI
Review profile: CHILL
Files selected for processing (1)
- metadata-integration/java/datahub-client/src/test/java/datahub/client/patch/PatchTest.java (4 hunks)
Files skipped from review as they are similar to previous changes (1)
- metadata-integration/java/datahub-client/src/test/java/datahub/client/patch/PatchTest.java
a7f4b71
into
datahub-project:master
* feat(forms) Handle deleting forms references when hard deleting forms (datahub-project#10820) * refactor(ui): Misc improvements to the setup ingestion flow (ingest uplift 1/2) (datahub-project#10764) Co-authored-by: John Joyce <[email protected]> Co-authored-by: John Joyce <[email protected]> * fix(ingestion/airflow-plugin): pipeline tasks discoverable in search (datahub-project#10819) * feat(ingest/transformer): tags to terms transformer (datahub-project#10758) Co-authored-by: Aseem Bansal <[email protected]> * fix(ingestion/unity-catalog): fixed issue with profiling with GE turned on (datahub-project#10752) Co-authored-by: Aseem Bansal <[email protected]> * feat(forms) Add java SDK for form entity PATCH + CRUD examples (datahub-project#10822) * feat(SDK) Add java SDK for structuredProperty entity PATCH + CRUD examples (datahub-project#10823) * feat(SDK) Add StructuredPropertyPatchBuilder in python sdk and provide sample CRUD files (datahub-project#10824) * feat(forms) Add CRUD endpoints to GraphQL for Form entities (datahub-project#10825) * add flag for includeSoftDeleted in scroll entities API (datahub-project#10831) * feat(deprecation) Return actor entity with deprecation aspect (datahub-project#10832) * feat(structuredProperties) Add CRUD graphql APIs for structured property entities (datahub-project#10826) * add scroll parameters to openapi v3 spec (datahub-project#10833) * fix(ingest): correct profile_day_of_week implementation (datahub-project#10818) * feat(ingest/glue): allow ingestion of empty databases from Glue (datahub-project#10666) Co-authored-by: Harshal Sheth <[email protected]> * feat(cli): add more details to get cli (datahub-project#10815) * fix(ingestion/glue): ensure date formatting works on all platforms for aws glue (datahub-project#10836) * fix(ingestion): fix datajob patcher (datahub-project#10827) * fix(smoke-test): add suffix in temp file creation (datahub-project#10841) * feat(ingest/glue): add helper method to permit user or group ownership (datahub-project#10784) * feat(): Show data platform instances in policy modal if they are set on the policy (datahub-project#10645) Co-authored-by: Hendrik Richert <[email protected]> * docs(patch): add patch documentation for how implementation works (datahub-project#10010) Co-authored-by: John Joyce <[email protected]> * fix(jar): add missing custom-plugin-jar task (datahub-project#10847) * fix(): also check exceptions/stack trace when filtering log messages (datahub-project#10391) Co-authored-by: John Joyce <[email protected]> * docs(): Update posts.md (datahub-project#9893) Co-authored-by: Hyejin Yoon <[email protected]> Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * chore(ingest): update acryl-datahub-classify version (datahub-project#10844) * refactor(ingest): Refactor structured logging to support infos, warnings, and failures structured reporting to UI (datahub-project#10828) Co-authored-by: John Joyce <[email protected]> Co-authored-by: Harshal Sheth <[email protected]> * fix(restli): log aspect-not-found as a warning rather than as an error (datahub-project#10834) * fix(ingest/nifi): remove duplicate upstream jobs (datahub-project#10849) * fix(smoke-test): test access to create/revoke personal access tokens (datahub-project#10848) * fix(smoke-test): missing test for move domain (datahub-project#10837) * ci: update usernames to not considered for community (datahub-project#10851) * env: change defaults for data contract visibility (datahub-project#10854) * fix(ingest/tableau): quote special characters in external URL (datahub-project#10842) * fix(smoke-test): fix flakiness of auto complete test * ci(ingest): pin dask dependency for feast (datahub-project#10865) * fix(ingestion/lookml): liquid template resolution and view-to-view cll (datahub-project#10542) * feat(ingest/audit): add client id and version in system metadata props (datahub-project#10829) * chore(ingest): Mypy 1.10.1 pin (datahub-project#10867) * docs: use acryl-datahub-actions as expected python package to install (datahub-project#10852) * docs: add new js snippet (datahub-project#10846) * refactor(ingestion): remove company domain for security reason (datahub-project#10839) * fix(ingestion/spark): Platform instance and column level lineage fix (datahub-project#10843) Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * feat(ingestion/tableau): optionally ingest multiple sites and create site containers (datahub-project#10498) Co-authored-by: Yanik Häni <[email protected]> * fix(ingestion/looker): Add sqlglot dependency and remove unused sqlparser (datahub-project#10874) * fix(manage-tokens): fix manage access token policy (datahub-project#10853) * Batch get entity endpoints (datahub-project#10880) * feat(system): support conditional write semantics (datahub-project#10868) * fix(build): upgrade vercel builds to Node 20.x (datahub-project#10890) * feat(ingest/lookml): shallow clone repos (datahub-project#10888) * fix(ingest/looker): add missing dependency (datahub-project#10876) * fix(ingest): only populate audit stamps where accurate (datahub-project#10604) * fix(ingest/dbt): always encode tag urns (datahub-project#10799) * fix(ingest/redshift): handle multiline alter table commands (datahub-project#10727) * fix(ingestion/looker): column name missing in explore (datahub-project#10892) * fix(lineage) Fix lineage source/dest filtering with explored per hop limit (datahub-project#10879) * feat(conditional-writes): misc updates and fixes (datahub-project#10901) * feat(ci): update outdated action (datahub-project#10899) * feat(rest-emitter): adding async flag to rest emitter (datahub-project#10902) Co-authored-by: Gabe Lyons <[email protected]> * feat(ingest): add snowflake-queries source (datahub-project#10835) * fix(ingest): improve `auto_materialize_referenced_tags_terms` error handling (datahub-project#10906) * docs: add new company to adoption list (datahub-project#10909) * refactor(redshift): Improve redshift error handling with new structured reporting system (datahub-project#10870) Co-authored-by: John Joyce <[email protected]> Co-authored-by: Harshal Sheth <[email protected]> * feat(ui) Finalize support for all entity types on forms (datahub-project#10915) * Index ExecutionRequestResults status field (datahub-project#10811) * feat(ingest): grafana connector (datahub-project#10891) Co-authored-by: Shirshanka Das <[email protected]> Co-authored-by: Harshal Sheth <[email protected]> * fix(gms) Add Form entity type to EntityTypeMapper (datahub-project#10916) * feat(dataset): add support for external url in Dataset (datahub-project#10877) * docs(saas-overview) added missing features to observe section (datahub-project#10913) Co-authored-by: John Joyce <[email protected]> * fix(ingest/spark): Fixing Micrometer warning (datahub-project#10882) * fix(structured properties): allow application of structured properties without schema file (datahub-project#10918) * fix(data-contracts-web) handle other schedule types (datahub-project#10919) * fix(ingestion/tableau): human-readable message for PERMISSIONS_MODE_SWITCHED error (datahub-project#10866) Co-authored-by: Harshal Sheth <[email protected]> * Add feature flag for view defintions (datahub-project#10914) Co-authored-by: Ethan Cartwright <[email protected]> * feat(ingest/BigQuery): refactor+parallelize dataset metadata extraction (datahub-project#10884) * fix(airflow): add error handling around render_template() (datahub-project#10907) * feat(ingestion/sqlglot): add optional `default_dialect` parameter to sqlglot lineage (datahub-project#10830) * feat(mcp-mutator): new mcp mutator plugin (datahub-project#10904) * fix(ingest/bigquery): changes helper function to decode unicode scape sequences (datahub-project#10845) * feat(ingest/postgres): fetch table sizes for profile (datahub-project#10864) * feat(ingest/abs): Adding azure blob storage ingestion source (datahub-project#10813) * fix(ingest/redshift): reduce severity of SQL parsing issues (datahub-project#10924) * fix(build): fix lint fix web react (datahub-project#10896) * fix(ingest/bigquery): handle quota exceeded for project.list requests (datahub-project#10912) * feat(ingest): report extractor failures more loudly (datahub-project#10908) * feat(ingest/snowflake): integrate snowflake-queries into main source (datahub-project#10905) * fix(ingest): fix docs build (datahub-project#10926) * fix(ingest/snowflake): fix test connection (datahub-project#10927) * fix(ingest/lookml): add view load failures to cache (datahub-project#10923) * docs(slack) overhauled setup instructions and screenshots (datahub-project#10922) Co-authored-by: John Joyce <[email protected]> * fix(airflow): Add comma parsing of owners to DataJobs (datahub-project#10903) * fix(entityservice): fix merging sideeffects (datahub-project#10937) * feat(ingest): Support System Ingestion Sources, Show and hide system ingestion sources with Command-S (datahub-project#10938) Co-authored-by: John Joyce <[email protected]> * chore() Set a default lineage filtering end time on backend when a start time is present (datahub-project#10925) Co-authored-by: John Joyce <[email protected]> Co-authored-by: John Joyce <[email protected]> * Added relationships APIs to V3. Added these generic APIs to V3 swagger doc. (datahub-project#10939) * docs: add learning center to docs (datahub-project#10921) * doc: Update hubspot form id (datahub-project#10943) * chore(airflow): add python 3.11 w/ Airflow 2.9 to CI (datahub-project#10941) * fix(ingest/Glue): column upstream lineage between S3 and Glue (datahub-project#10895) * fix(ingest/abs): split abs utils into multiple files (datahub-project#10945) * doc(ingest/looker): fix doc for sql parsing documentation (datahub-project#10883) Co-authored-by: Harshal Sheth <[email protected]> * fix(ingest/bigquery): Adding missing BigQuery types (datahub-project#10950) * fix(ingest/setup): feast and abs source setup (datahub-project#10951) * fix(connections) Harden adding /gms to connections in backend (datahub-project#10942) * feat(siblings) Add flag to prevent combining siblings in the UI (datahub-project#10952) * fix(docs): make graphql doc gen more automated (datahub-project#10953) * feat(ingest/athena): Add option for Athena partitioned profiling (datahub-project#10723) * fix(spark-lineage): default timeout for future responses (datahub-project#10947) * feat(datajob/flow): add environment filter using info aspects (datahub-project#10814) * fix(ui/ingest): correct privilege used to show tab (datahub-project#10483) Co-authored-by: Kunal-kankriya <[email protected]> * feat(ingest/looker): include dashboard urns in browse v2 (datahub-project#10955) * add a structured type to batchGet in OpenAPI V3 spec (datahub-project#10956) * fix(ui): scroll on the domain sidebar to show all domains (datahub-project#10966) * fix(ingest/sagemaker): resolve incorrect variable assignment for SageMaker API call (datahub-project#10965) * fix(airflow/build): Pinning mypy (datahub-project#10972) * Fixed a bug where the OpenAPI V3 spec was incorrect. The bug was introduced in datahub-project#10939. (datahub-project#10974) * fix(ingest/test): Fix for mssql integration tests (datahub-project#10978) * fix(entity-service) exist check correctly extracts status (datahub-project#10973) * fix(structuredProps) casing bug in StructuredPropertiesValidator (datahub-project#10982) * bugfix: use anyOf instead of allOf when creating references in openapi v3 spec (datahub-project#10986) * fix(ui): Remove ant less imports (datahub-project#10988) * feat(ingest/graph): Add get_results_by_filter to DataHubGraph (datahub-project#10987) * feat(ingest/cli): init does not actually support environment variables (datahub-project#10989) * fix(ingest/graph): Update get_results_by_filter graphql query (datahub-project#10991) * feat(ingest/spark): Promote beta plugin (datahub-project#10881) Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * feat(ingest): support domains in meta -> "datahub" section (datahub-project#10967) * feat(ingest): add `check server-config` command (datahub-project#10990) * feat(cli): Make consistent use of DataHubGraphClientConfig (datahub-project#10466) Deprecates get_url_and_token() in favor of a more complete option: load_graph_config() that returns a full DatahubClientConfig. This change was then propagated across previous usages of get_url_and_token so that connections to DataHub server from the client respect the full breadth of configuration specified by DatahubClientConfig. I.e: You can now specify disable_ssl_verification: true in your ~/.datahubenv file so that all cli functions to the server work when ssl certification is disabled. Fixes datahub-project#9705 * fix(ingest/s3): Fixing container creation when there is no folder in path (datahub-project#10993) * fix(ingest/looker): support platform instance for dashboards & charts (datahub-project#10771) * feat(ingest/bigquery): improve handling of information schema in sql parser (datahub-project#10985) * feat(ingest): improve `ingest deploy` command (datahub-project#10944) * fix(backend): allow excluding soft-deleted entities in relationship-queries; exclude soft-deleted members of groups (datahub-project#10920) - allow excluding soft-deleted entities in relationship-queries - exclude soft-deleted members of groups * fix(ingest/looker): downgrade missing chart type log level (datahub-project#10996) * doc(acryl-cloud): release docs for 0.3.4.x (datahub-project#10984) Co-authored-by: John Joyce <[email protected]> Co-authored-by: RyanHolstien <[email protected]> Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> Co-authored-by: Pedro Silva <[email protected]> * fix(protobuf/build): Fix protobuf check jar script (datahub-project#11006) * fix(ui/ingest): Support invalid cron jobs (datahub-project#10998) * fix(ingest): fix graph config loading (datahub-project#11002) Co-authored-by: Pedro Silva <[email protected]> * feat(docs): Document __DATAHUB_TO_FILE_ directive (datahub-project#10968) Co-authored-by: Harshal Sheth <[email protected]> * fix(graphql/upsertIngestionSource): Validate cron schedule; parse error in CLI (datahub-project#11011) * feat(ece): support custom ownership type urns in ECE generation (datahub-project#10999) * feat(assertion-v2): changed Validation tab to Quality and created new Governance tab (datahub-project#10935) * fix(ingestion/glue): Add support for missing config options for profiling in Glue (datahub-project#10858) * feat(propagation): Add models for schema field docs, tags, terms (datahub-project#2959) (datahub-project#11016) Co-authored-by: Chris Collins <[email protected]> * docs: standardize terminology to DataHub Cloud (datahub-project#11003) * fix(ingestion/transformer): replace the externalUrl container (datahub-project#11013) * docs(slack) troubleshoot docs (datahub-project#11014) * feat(propagation): Add graphql API (datahub-project#11030) Co-authored-by: Chris Collins <[email protected]> * feat(propagation): Add models for Action feature settings (datahub-project#11029) * docs(custom properties): Remove duplicate from sidebar (datahub-project#11033) * feat(models): Introducing Dataset Partitions Aspect (datahub-project#10997) Co-authored-by: John Joyce <[email protected]> Co-authored-by: John Joyce <[email protected]> * feat(propagation): Add Documentation Propagation Settings (datahub-project#11038) * fix(models): chart schema fields mapping, add dataHubAction entity, t… (datahub-project#11040) * fix(ci): smoke test lint failures (datahub-project#11044) * docs: fix learning center color scheme & typo (datahub-project#11043) * feat: add cloud main page (datahub-project#11017) Co-authored-by: Jay <[email protected]> * feat(restore-indices): add additional step to also clear system metadata service (datahub-project#10662) Co-authored-by: John Joyce <[email protected]> * docs: fix typo (datahub-project#11046) * fix(lint): apply spotless (datahub-project#11050) * docs(airflow): example query to get datajobs for a dataflow (datahub-project#11034) * feat(cli): Add run-id option to put sub-command (datahub-project#11023) Adds an option to assign run-id to a given put command execution. This is useful when transformers do not exist for a given ingestion payload, we can follow up with custom metadata and assign it to an ingestion pipeline. * fix(ingest): improve sql error reporting calls (datahub-project#11025) * fix(airflow): fix CI setup (datahub-project#11031) * feat(ingest/dbt): add experimental `prefer_sql_parser_lineage` flag (datahub-project#11039) * fix(ingestion/lookml): enable stack-trace in lookml logs (datahub-project#10971) * (chore): Linting fix (datahub-project#11015) * chore(ci): update deprecated github actions (datahub-project#10977) * Fix ALB configuration example (datahub-project#10981) * chore(ingestion-base): bump base image packages (datahub-project#11053) * feat(cli): Trim report of dataHubExecutionRequestResult to max GMS size (datahub-project#11051) * fix(ingestion/lookml): emit dummy sql condition for lookml custom condition tag (datahub-project#11008) Co-authored-by: Harshal Sheth <[email protected]> * fix(ingestion/powerbi): fix issue with broken report lineage (datahub-project#10910) * feat(ingest/tableau): add retry on timeout (datahub-project#10995) * change generate kafka connect properties from env (datahub-project#10545) Co-authored-by: david-leifker <[email protected]> * fix(ingest): fix oracle cronjob ingestion (datahub-project#11001) Co-authored-by: david-leifker <[email protected]> * chore(ci): revert update deprecated github actions (datahub-project#10977) (datahub-project#11062) * feat(ingest/dbt-cloud): update metadata_endpoint inference (datahub-project#11041) * build: Reduce size of datahub-frontend-react image by 50-ish% (datahub-project#10878) Co-authored-by: david-leifker <[email protected]> * fix(ci): Fix lint issue in datahub_ingestion_run_summary_provider.py (datahub-project#11063) * docs(ingest): update developing-a-transformer.md (datahub-project#11019) * feat(search-test): update search tests from datahub-project#10408 (datahub-project#11056) * feat(cli): add aspects parameter to DataHubGraph.get_entity_semityped (datahub-project#11009) Co-authored-by: Harshal Sheth <[email protected]> * docs(airflow): update min version for plugin v2 (datahub-project#11065) * doc(ingestion/tableau): doc update for derived permission (datahub-project#11054) Co-authored-by: Pedro Silva <[email protected]> Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> Co-authored-by: Harshal Sheth <[email protected]> * fix(py): remove dep on types-pkg_resources (datahub-project#11076) * feat(ingest/mode): add option to exclude restricted (datahub-project#11081) * fix(ingest): set lastObserved in sdk when unset (datahub-project#11071) * doc(ingest): Update capabilities (datahub-project#11072) * chore(vulnerability): Log Injection (datahub-project#11090) * chore(vulnerability): Information exposure through a stack trace (datahub-project#11091) * chore(vulnerability): Comparison of narrow type with wide type in loop condition (datahub-project#11089) * chore(vulnerability): Insertion of sensitive information into log files (datahub-project#11088) * chore(vulnerability): Risky Cryptographic Algorithm (datahub-project#11059) * chore(vulnerability): Overly permissive regex range (datahub-project#11061) Co-authored-by: Harshal Sheth <[email protected]> * fix: update customer data (datahub-project#11075) * fix(models): fixing the datasetPartition models (datahub-project#11085) Co-authored-by: John Joyce <[email protected]> * fix(ui): Adding view, forms GraphQL query, remove showing a fallback error message on unhandled GraphQL error (datahub-project#11084) Co-authored-by: John Joyce <[email protected]> * feat(docs-site): hiding learn more from cloud page (datahub-project#11097) * fix(docs): Add correct usage of orFilters in search API docs (datahub-project#11082) Co-authored-by: Jay <[email protected]> * fix(ingest/mode): Regexp in mode name matcher didn't allow underscore (datahub-project#11098) * docs: Refactor customer stories section (datahub-project#10869) Co-authored-by: Jeff Merrick <[email protected]> * fix(release): fix full/slim suffix on tag (datahub-project#11087) * feat(config): support alternate hashing algorithm for doc id (datahub-project#10423) Co-authored-by: david-leifker <[email protected]> Co-authored-by: John Joyce <[email protected]> * fix(emitter): fix typo in get method of java kafka emitter (datahub-project#11007) * fix(ingest): use correct native data type in all SQLAlchemy sources by compiling data type using dialect (datahub-project#10898) Co-authored-by: Harshal Sheth <[email protected]> * chore: Update contributors list in PR labeler (datahub-project#11105) * feat(ingest): tweak stale entity removal messaging (datahub-project#11064) * fix(ingestion): enforce lastObserved timestamps in SystemMetadata (datahub-project#11104) * fix(ingest/powerbi): fix broken lineage between chart and dataset (datahub-project#11080) * feat(ingest/lookml): CLL support for sql set in sql_table_name attribute of lookml view (datahub-project#11069) * docs: update graphql docs on forms & structured properties (datahub-project#11100) * test(search): search openAPI v3 test (datahub-project#11049) * fix(ingest/tableau): prevent empty site content urls (datahub-project#11057) Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> * feat(entity-client): implement client batch interface (datahub-project#11106) * fix(snowflake): avoid reporting warnings/info for sys tables (datahub-project#11114) * fix(ingest): downgrade column type mapping warning to info (datahub-project#11115) * feat(api): add AuditStamp to the V3 API entity/aspect response (datahub-project#11118) * fix(ingest/redshift): replace r'\n' with '\n' to avoid token error redshift serverless… (datahub-project#11111) * fix(entiy-client): handle null entityUrn case for restli (datahub-project#11122) * fix(sql-parser): prevent bad urns from alter table lineage (datahub-project#11092) * fix(ingest/bigquery): use small batch size if use_tables_list_query_v2 is set (datahub-project#11121) * fix(graphql): add missing entities to EntityTypeMapper and EntityTypeUrnMapper (datahub-project#10366) * feat(ui): Changes to allow editable dataset name (datahub-project#10608) Co-authored-by: Jay Kadambi <[email protected]> * fix: remove saxo (datahub-project#11127) * feat(mcl-processor): Update mcl processor hooks (datahub-project#11134) * fix(openapi): fix openapi v2 endpoints & v3 documentation update * Revert "fix(openapi): fix openapi v2 endpoints & v3 documentation update" This reverts commit 573c1cb. * docs(policies): updates to policies documentation (datahub-project#11073) * fix(openapi): fix openapi v2 and v3 docs update (datahub-project#11139) * feat(auth): grant type and acr values custom oidc parameters support (datahub-project#11116) * fix(mutator): mutator hook fixes (datahub-project#11140) * feat(search): support sorting on multiple fields (datahub-project#10775) * feat(ingest): various logging improvements (datahub-project#11126) * fix(ingestion/lookml): fix for sql parsing error (datahub-project#11079) Co-authored-by: Harshal Sheth <[email protected]> * feat(docs-site) cloud page spacing and content polishes (datahub-project#11141) * feat(ui) Enable editing structured props on fields (datahub-project#11042) * feat(tests): add md5 and last computed to testResult model (datahub-project#11117) * test(openapi): openapi regression smoke tests (datahub-project#11143) * fix(airflow): fix tox tests + update docs (datahub-project#11125) * docs: add chime to adoption stories (datahub-project#11142) * fix(ingest/databricks): Updating code to work with Databricks sdk 0.30 (datahub-project#11158) * fix(kafka-setup): add missing script to image (datahub-project#11190) * fix(config): fix hash algo config (datahub-project#11191) * test(smoke-test): updates to smoke-tests (datahub-project#11152) * fix(elasticsearch): refactor idHashAlgo setting (datahub-project#11193) * chore(kafka): kafka version bump (datahub-project#11211) * readd UsageStatsWorkUnit * fix merge problems * change logo --------- Co-authored-by: Chris Collins <[email protected]> Co-authored-by: John Joyce <[email protected]> Co-authored-by: John Joyce <[email protected]> Co-authored-by: John Joyce <[email protected]> Co-authored-by: dushayntAW <[email protected]> Co-authored-by: sagar-salvi-apptware <[email protected]> Co-authored-by: Aseem Bansal <[email protected]> Co-authored-by: Kevin Chun <[email protected]> Co-authored-by: jordanjeremy <[email protected]> Co-authored-by: skrydal <[email protected]> Co-authored-by: Harshal Sheth <[email protected]> Co-authored-by: david-leifker <[email protected]> Co-authored-by: sid-acryl <[email protected]> Co-authored-by: Julien Jehannet <[email protected]> Co-authored-by: Hendrik Richert <[email protected]> Co-authored-by: Hendrik Richert <[email protected]> Co-authored-by: RyanHolstien <[email protected]> Co-authored-by: Felix Lüdin <[email protected]> Co-authored-by: Pirry <[email protected]> Co-authored-by: Hyejin Yoon <[email protected]> Co-authored-by: coderabbitai[bot] <136622811+coderabbitai[bot]@users.noreply.github.com> Co-authored-by: cburroughs <[email protected]> Co-authored-by: ksrinath <[email protected]> Co-authored-by: Mayuri Nehate <[email protected]> Co-authored-by: Kunal-kankriya <[email protected]> Co-authored-by: Shirshanka Das <[email protected]> Co-authored-by: ipolding-cais <[email protected]> Co-authored-by: Tamas Nemeth <[email protected]> Co-authored-by: Shubham Jagtap <[email protected]> Co-authored-by: haeniya <[email protected]> Co-authored-by: Yanik Häni <[email protected]> Co-authored-by: Gabe Lyons <[email protected]> Co-authored-by: Gabe Lyons <[email protected]> Co-authored-by: 808OVADOZE <[email protected]> Co-authored-by: noggi <[email protected]> Co-authored-by: Nicholas Pena <[email protected]> Co-authored-by: Jay <[email protected]> Co-authored-by: ethan-cartwright <[email protected]> Co-authored-by: Ethan Cartwright <[email protected]> Co-authored-by: Nadav Gross <[email protected]> Co-authored-by: Patrick Franco Braz <[email protected]> Co-authored-by: pie1nthesky <[email protected]> Co-authored-by: Joel Pinto Mata (KPN-DSH-DEX team) <[email protected]> Co-authored-by: Ellie O'Neil <[email protected]> Co-authored-by: Ajoy Majumdar <[email protected]> Co-authored-by: deepgarg-visa <[email protected]> Co-authored-by: Tristan Heisler <[email protected]> Co-authored-by: Andrew Sikowitz <[email protected]> Co-authored-by: Davi Arnaut <[email protected]> Co-authored-by: Pedro Silva <[email protected]> Co-authored-by: amit-apptware <[email protected]> Co-authored-by: Sam Black <[email protected]> Co-authored-by: Raj Tekal <[email protected]> Co-authored-by: Steffen Grohsschmiedt <[email protected]> Co-authored-by: jaegwon.seo <[email protected]> Co-authored-by: Renan F. Lima <[email protected]> Co-authored-by: Matt Exchange <[email protected]> Co-authored-by: Jonny Dixon <[email protected]> Co-authored-by: Pedro Silva <[email protected]> Co-authored-by: Pinaki Bhattacharjee <[email protected]> Co-authored-by: Jeff Merrick <[email protected]> Co-authored-by: skrydal <[email protected]> Co-authored-by: AndreasHegerNuritas <[email protected]> Co-authored-by: jayasimhankv <[email protected]> Co-authored-by: Jay Kadambi <[email protected]> Co-authored-by: David Leifker <[email protected]>
This PR tackles two main pieces in the Java SDK around structured properties:
So now you should be able to create structured properties and then assign/patch them on assets using the java SDK.
Checklist
Summary by CodeRabbit
New Features
Tests
These updates enhance the flexibility and capabilities of managing structured properties, providing users with robust options for handling complex data structures.