You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Dec 20, 2018. It is now read-only.
I cannot find RddUtils.rddToDataFrame. Where is this api? Basically, I want to pass bytes[] and avroSchema json string. Given these two - can I get a GenericRecord back?
The text was updated successfully, but these errors were encountered:
@gsolasab, @squito, @gengliangwang@rxin
Can some one look in to this request. We have a use case to write AVRO records to Kafka.
We read some data from Kafka as String/Ascii, perform some ETL operation like extract zipped content, join on some lookup and finally prepare some nested row DataFrame. Finally write to Kafka as AVRO.
In this process I am looking for generic way to convert Row to avro byte[] so that I can write it to Kafka.
Hi
I have a use case to convert a Row to AVRO and write those Bytes to Kafka Sink. Kindly help me.
I tried this:
Tried This:
This works for Flat AVRO schemas. If we have nested, it fails.
Write To Kafka: (Need Help)
I am expecting an API like:
If you can expose this private method - that will do it I guess
https://github.com/databricks/spark-avro/blob/branch-4.0/src/main/scala/com/databricks/spark/avro/AvroOutputWriter.scala-->converter(row).asInstanceOf[GenericRecord])
Read from Kafka:
I cannot find RddUtils.rddToDataFrame. Where is this api? Basically, I want to pass bytes[] and avroSchema json string. Given these two - can I get a GenericRecord back?
The text was updated successfully, but these errors were encountered: