当前位置: 首页>>代码示例>>Java>>正文


Java RestClientException类代码示例

本文整理汇总了Java中io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException的典型用法代码示例。如果您正苦于以下问题:Java RestClientException类的具体用法?Java RestClientException怎么用?Java RestClientException使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。


RestClientException类属于io.confluent.kafka.schemaregistry.client.rest.exceptions包,在下文中一共展示了RestClientException类的14个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的Java代码示例。

示例1: registerSchema

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
/**
 * Register the schema with the kafka schema registry.
 * @param schemaName the name of the schema
 * @param schema the avro schema
 * @return the id for the schema returned from the schema registry.
 */
public int registerSchema(String schemaName, Schema schema) {
    int id;
    
    try {
        /*
         * "-value" appended to schema name per Kafka documentation.
         * See link: : http://docs.confluent.io/2.0.0/schema-registry/docs/api.html
         */
        id = client.register(schemaName+"-value", schema);
    } catch (RestClientException | IOException e) {
        LOG.error("registerSchema(): failed. Returning ID=0", e);
        id = 0;
    }

    LOG.debug("RegisterSchema(): Schema ID returned from registry for {} was {}", schemaName, id);
    
    return id;
}
 
开发者ID:oracle,项目名称:bdglue,代码行数:25,代码来源:KafkaRegistry.java

示例2: fetchSchemaMetadata

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
private SchemaMetadata fetchSchemaMetadata(
    AbstractStreamCreateStatement abstractStreamCreateStatement,
    SchemaRegistryClient schemaRegistryClient,
    String kafkaTopicName) throws IOException, RestClientException {

  if (abstractStreamCreateStatement.getProperties().containsKey(KsqlConstants.AVRO_SCHEMA_ID)) {
    int schemaId;
    try {
      schemaId = Integer.parseInt(StringUtil.cleanQuotes(abstractStreamCreateStatement.getProperties().get(KsqlConstants.AVRO_SCHEMA_ID).toString()));
    } catch (NumberFormatException e) {
      throw new KsqlException(String.format("Invalid schema id property: %s.",
                                            abstractStreamCreateStatement.getProperties().get(KsqlConstants.AVRO_SCHEMA_ID).toString()));
    }
    return schemaRegistryClient.getSchemaMetadata(
        kafkaTopicName +KsqlConstants.SCHEMA_REGISTRY_VALUE_SUFFIX, schemaId);
  } else {
    return schemaRegistryClient.getLatestSchemaMetadata(kafkaTopicName +
                                                                  KsqlConstants.SCHEMA_REGISTRY_VALUE_SUFFIX);
  }
}
 
开发者ID:confluentinc,项目名称:ksql,代码行数:21,代码来源:AvroUtil.java

示例3: shouldValidatePersistentQueryResultCorrectly

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Test
public void shouldValidatePersistentQueryResultCorrectly()
    throws IOException, RestClientException {
  SchemaRegistryClient schemaRegistryClient = mock(SchemaRegistryClient.class);
  KsqlTopic resultTopic = new KsqlTopic("testTopic", "testTopic", new KsqlAvroTopicSerDe());
  Schema resultSchema = SerDeUtil.getSchemaFromAvro(ordersAveroSchemaStr);
  PersistentQueryMetadata persistentQueryMetadata = new PersistentQueryMetadata("",
                                                                                null,
                                                                                null,
                                                                                "",
                                                                                null,
                                                                                DataSource.DataSourceType.KSTREAM,
                                                                                "",
                                                                                mock(KafkaTopicClient.class),
      resultSchema,
                                                                                resultTopic,
                                                                                null);
  org.apache.avro.Schema.Parser parser = new org.apache.avro.Schema.Parser();
  org.apache.avro.Schema avroSchema = parser.parse(ordersAveroSchemaStr);
  expect(schemaRegistryClient.testCompatibility(anyString(), EasyMock.isA(avroSchema.getClass())))
      .andReturn(true);
  replay(schemaRegistryClient);
  avroUtil.validatePersistentQueryResults(persistentQueryMetadata, schemaRegistryClient);
}
 
开发者ID:confluentinc,项目名称:ksql,代码行数:25,代码来源:AvroUtilTest.java

示例4: registerSchema

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
private void registerSchema(SchemaRegistryClient schemaRegistryClient)
    throws IOException, RestClientException {
  String ordersAveroSchemaStr = "{"
                                + "\"namespace\": \"kql\","
                                + " \"name\": \"orders\","
                                + " \"type\": \"record\","
                                + " \"fields\": ["
                                + "     {\"name\": \"ordertime\", \"type\": \"long\"},"
                                + "     {\"name\": \"orderid\",  \"type\": \"long\"},"
                                + "     {\"name\": \"itemid\", \"type\": \"string\"},"
                                + "     {\"name\": \"orderunits\", \"type\": \"double\"},"
                                + "     {\"name\": \"arraycol\", \"type\": {\"type\": \"array\", \"items\": \"double\"}},"
                                + "     {\"name\": \"mapcol\", \"type\": {\"type\": \"map\", \"values\": \"double\"}}"
                                + " ]"
                                + "}";
  org.apache.avro.Schema.Parser parser = new org.apache.avro.Schema.Parser();
  org.apache.avro.Schema avroSchema = parser.parse(ordersAveroSchemaStr);
  schemaRegistryClient.register("orders-topic" + KsqlConstants.SCHEMA_REGISTRY_VALUE_SUFFIX,
                                avroSchema);

}
 
开发者ID:confluentinc,项目名称:ksql,代码行数:22,代码来源:KsqlResourceTest.java

示例5: retrieveSchema

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Override
public Schema retrieveSchema(TableId table, String topic) {
  try {
    String subject = getSubject(topic);
    logger.debug("Retrieving schema information for topic {} with subject {}", topic, subject);
    SchemaMetadata latestSchemaMetadata = schemaRegistryClient.getLatestSchemaMetadata(subject);
    org.apache.avro.Schema avroSchema = new Parser().parse(latestSchemaMetadata.getSchema());
    return avroData.toConnectSchema(avroSchema);
  } catch (IOException | RestClientException exception) {
    throw new ConnectException(
        "Exception encountered while trying to fetch latest schema metadata from Schema Registry",
        exception
    );
  }
}
 
开发者ID:wepay,项目名称:kafka-connect-bigquery,代码行数:16,代码来源:SchemaRegistrySchemaRetriever.java

示例6: shouldFailForInvalidResultAvroSchema

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Test
public void shouldFailForInvalidResultAvroSchema()
    throws IOException, RestClientException {
  SchemaRegistryClient schemaRegistryClient = mock(SchemaRegistryClient.class);
  KsqlTopic resultTopic = new KsqlTopic("testTopic", "testTopic", new KsqlAvroTopicSerDe
      ());
  Schema resultSchema = SerDeUtil.getSchemaFromAvro(ordersAveroSchemaStr);
  PersistentQueryMetadata persistentQueryMetadata = new PersistentQueryMetadata("",
                                                                                null,
                                                                                null,
                                                                                "",
                                                                                null,
                                                                                DataSource.DataSourceType.KSTREAM,
                                                                                "",
                                                                                mock(KafkaTopicClient.class),
      resultSchema,
                                                                                resultTopic,
                                                                                null);
  expect(schemaRegistryClient.testCompatibility(anyString(), anyObject())).andReturn(false);
  replay(schemaRegistryClient);
  try {
    avroUtil.validatePersistentQueryResults(persistentQueryMetadata, schemaRegistryClient);
    fail();
  } catch (Exception e) {
    assertThat("Incorrect exception message", "Cannot register avro schema for testTopic since "
                                              + "it is not valid for schema registry.", equalTo(e.getMessage()));
  }
}
 
开发者ID:confluentinc,项目名称:ksql,代码行数:29,代码来源:AvroUtilTest.java

示例7: setUp

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Before
public void setUp() throws IOException, RestClientException {
  schemaRegistryClient = new MockSchemaRegistryClient();
  registerSchema(schemaRegistryClient);
  ksqlRestConfig = new KsqlRestConfig(TestKsqlResourceUtil.getDefaultKsqlConfig());
  KsqlConfig ksqlConfig = new KsqlConfig(ksqlRestConfig.getKsqlStreamsProperties());
  ksqlEngine = new KsqlEngine(ksqlConfig, new MockKafkaTopicClient(), schemaRegistryClient);
}
 
开发者ID:confluentinc,项目名称:ksql,代码行数:9,代码来源:KsqlResourceTest.java

示例8: loadFromRegistry

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
/**
 * Loads and parses a schema for the specified subject from the schema registry
 * @param subject subject for which to fetch the latest version of a schema.
 * @return parsed avro schema
 * @throws SchemaRegistryException if there is an error with the registry client
 */
public Schema loadFromRegistry(String subject) throws SchemaRegistryException {
  try {
    SchemaMetadata metadata = registryClient.getLatestSchemaMetadata(subject);
    return registryClient.getByID(metadata.getId());
  } catch (IOException | RestClientException e) {
    throw new SchemaRegistryException(e);
  }
}
 
开发者ID:streamsets,项目名称:datacollector,代码行数:15,代码来源:AvroSchemaHelper.java

示例9: getSchemaIdFromSubject

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
/**
 * Looks up schema id for the specified subject from the schema registry
 * @param subject subject for which schema Id must be looked up.
 * @return the schema id
 * @throws SchemaRegistryException if there is an error with the registry client
 */
public int getSchemaIdFromSubject(String subject) throws SchemaRegistryException {
  try {
    SchemaMetadata metadata = registryClient.getLatestSchemaMetadata(subject);
    return metadata.getId();
  } catch (IOException | RestClientException e) {
    throw new SchemaRegistryException(e);
  }
}
 
开发者ID:streamsets,项目名称:datacollector,代码行数:15,代码来源:AvroSchemaHelper.java

示例10: fetchSchemaByKey

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Override
protected Schema fetchSchemaByKey(Integer key) throws SchemaRegistryException {
  try {
    return this.schemaRegistryClient.getByID(key);
  } catch (IOException | RestClientException e) {
    throw new SchemaRegistryException(e);
  }
}
 
开发者ID:apache,项目名称:incubator-gobblin,代码行数:9,代码来源:ConfluentKafkaSchemaRegistry.java

示例11: getLatestSchemaByTopic

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Override
public Schema getLatestSchemaByTopic(String topic) throws SchemaRegistryException {
  String schemaName = topic + this.schemaNameSuffix;
  try {
    return new Schema.Parser().parse(this.schemaRegistryClient.getLatestSchemaMetadata(schemaName).getSchema());
  } catch (IOException | RestClientException e) {
    log.error("Failed to get schema for topic " + topic + "; subject " + schemaName);
    throw new SchemaRegistryException(e);
  }
}
 
开发者ID:apache,项目名称:incubator-gobblin,代码行数:11,代码来源:ConfluentKafkaSchemaRegistry.java

示例12: register

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Override
public Integer register(Schema schema, String name) throws SchemaRegistryException {
  try {
    String schemaName = name + this.schemaNameSuffix;
    return this.schemaRegistryClient.register(schemaName, schema);
  } catch (IOException | RestClientException e) {
    throw new SchemaRegistryException(e);
  }
}
 
开发者ID:apache,项目名称:incubator-gobblin,代码行数:10,代码来源:ConfluentKafkaSchemaRegistry.java

示例13: testConfluentAvroDeserializer

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Test
public void testConfluentAvroDeserializer() throws IOException, RestClientException {
  WorkUnitState mockWorkUnitState = getMockWorkUnitState();
  mockWorkUnitState.setProp("schema.registry.url", TEST_URL);

  Schema schema = SchemaBuilder.record(TEST_RECORD_NAME)
      .namespace(TEST_NAMESPACE).fields()
      .name(TEST_FIELD_NAME).type().stringType().noDefault()
      .endRecord();

  GenericRecord testGenericRecord = new GenericRecordBuilder(schema).set(TEST_FIELD_NAME, "testValue").build();

  SchemaRegistryClient mockSchemaRegistryClient = mock(SchemaRegistryClient.class);
  when(mockSchemaRegistryClient.getByID(any(Integer.class))).thenReturn(schema);

  Serializer<Object> kafkaEncoder = new KafkaAvroSerializer(mockSchemaRegistryClient);
  Deserializer<Object> kafkaDecoder = new KafkaAvroDeserializer(mockSchemaRegistryClient);

  ByteBuffer testGenericRecordByteBuffer =
      ByteBuffer.wrap(kafkaEncoder.serialize(TEST_TOPIC_NAME, testGenericRecord));

  KafkaSchemaRegistry<Integer, Schema> mockKafkaSchemaRegistry = mock(KafkaSchemaRegistry.class);
  KafkaDeserializerExtractor kafkaDecoderExtractor =
      new KafkaDeserializerExtractor(mockWorkUnitState,
          Optional.fromNullable(Deserializers.CONFLUENT_AVRO), kafkaDecoder, mockKafkaSchemaRegistry);

  ByteArrayBasedKafkaRecord mockMessageAndOffset = getMockMessageAndOffset(testGenericRecordByteBuffer);

  Assert.assertEquals(kafkaDecoderExtractor.decodeRecord(mockMessageAndOffset), testGenericRecord);
}
 
开发者ID:apache,项目名称:incubator-gobblin,代码行数:31,代码来源:KafkaDeserializerExtractorTest.java

示例14: testConfluentAvroDeserializerForSchemaEvolution

import io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException; //导入依赖的package包/类
@Test
public void testConfluentAvroDeserializerForSchemaEvolution() throws IOException, RestClientException, SchemaRegistryException {
  WorkUnitState mockWorkUnitState = getMockWorkUnitState();
  mockWorkUnitState.setProp("schema.registry.url", TEST_URL);

  Schema schemaV1 = SchemaBuilder.record(TEST_RECORD_NAME)
      .namespace(TEST_NAMESPACE).fields()
      .name(TEST_FIELD_NAME).type().stringType().noDefault()
      .endRecord();

  Schema schemaV2 = SchemaBuilder.record(TEST_RECORD_NAME)
      .namespace(TEST_NAMESPACE).fields()
      .name(TEST_FIELD_NAME).type().stringType().noDefault()
      .optionalString(TEST_FIELD_NAME2).endRecord();

  GenericRecord testGenericRecord = new GenericRecordBuilder(schemaV1).set(TEST_FIELD_NAME, "testValue").build();

  SchemaRegistryClient mockSchemaRegistryClient = mock(SchemaRegistryClient.class);
  when(mockSchemaRegistryClient.getByID(any(Integer.class))).thenReturn(schemaV1);

  Serializer<Object> kafkaEncoder = new KafkaAvroSerializer(mockSchemaRegistryClient);
  Deserializer<Object> kafkaDecoder = new KafkaAvroDeserializer(mockSchemaRegistryClient);

  ByteBuffer testGenericRecordByteBuffer =
      ByteBuffer.wrap(kafkaEncoder.serialize(TEST_TOPIC_NAME, testGenericRecord));

  KafkaSchemaRegistry<Integer, Schema> mockKafkaSchemaRegistry = mock(KafkaSchemaRegistry.class);
  when(mockKafkaSchemaRegistry.getLatestSchemaByTopic(TEST_TOPIC_NAME)).thenReturn(schemaV2);

  KafkaDeserializerExtractor kafkaDecoderExtractor = new KafkaDeserializerExtractor(mockWorkUnitState,
      Optional.fromNullable(Deserializers.CONFLUENT_AVRO), kafkaDecoder, mockKafkaSchemaRegistry);
  when(kafkaDecoderExtractor.getSchema()).thenReturn(schemaV2);

  ByteArrayBasedKafkaRecord mockMessageAndOffset = getMockMessageAndOffset(testGenericRecordByteBuffer);

  GenericRecord received = (GenericRecord) kafkaDecoderExtractor.decodeRecord(mockMessageAndOffset);
  Assert.assertEquals(received.toString(), "{\"testField\": \"testValue\", \"testField2\": null}");

}
 
开发者ID:apache,项目名称:incubator-gobblin,代码行数:40,代码来源:KafkaDeserializerExtractorTest.java


注:本文中的io.confluent.kafka.schemaregistry.client.rest.exceptions.RestClientException类示例由纯净天空整理自Github/MSDocs等开源代码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。