Encode and decode Avro binary data.
Parse an Avro JSON schema, then encode and decode Go values directly — no code generation required. Supports all primitive and complex types, logical types, schema evolution, Object Container Files, Single Object Encoding, and fingerprinting.
- Quick Start
- Type Mapping
- Struct Tags
- Schema Inference
- Schema Introspection
- Logical Types
- Schema Evolution
- Schema Cache
- Custom Types
- Object Container Files
- JSON Encoding
- Single Object Encoding
- Fingerprinting
- Performance
package main
import (
"fmt"
"log"
"github.com/twmb/avro"
)
var schema = avro.MustParse(`{
"type": "record",
"name": "User",
"fields": [
{"name": "name", "type": "string"},
{"name": "age", "type": "int"}
]
}`)
type User struct {
Name string `avro:"name"`
Age int `avro:"age"`
}
func main() {
// Encode
data, err := schema.Encode(&User{Name: "Alice", Age: 30})
if err != nil {
log.Fatal(err)
}
// Decode
var u User
_, err = schema.Decode(data, &u)
if err != nil {
log.Fatal(err)
}
fmt.Println(u) // {Alice 30}
}Parse accepts options: pass WithLaxNames() to allow non-standard characters
in type and field names (useful for interop with schemas from other languages).
The table below shows which Go types can be used with each Avro type.
| Avro Type | Encode | Decode |
|---|---|---|
| null | any (nil) |
any |
| boolean | bool |
bool, any |
| int, long | int, int8–int64, uint–uint64, float64, json.Number |
int, int8–int64, uint–uint64, any |
| float | float32, float64, json.Number |
float32, float64, any |
| double | float64, float32, json.Number |
float64, float32, any |
| string | string, []byte, encoding.TextAppender, encoding.TextMarshaler |
string, []byte, encoding.TextUnmarshaler, any |
| bytes | []byte, string |
[]byte, string, any |
| enum | string, any integer type (ordinal) |
string, any integer type (ordinal), any |
| fixed | [N]byte, []byte |
[N]byte, []byte, any |
| array | slice | slice, any |
| map | map[string]T |
map[string]T, any |
| union | any, *T, or the matched branch type |
any, *T, or the matched branch type |
| record | struct, map[string]any |
struct, map[string]any, any |
When decoding into any, values use their natural Go types: nil, bool,
int32, int64, float32, float64, string, []byte, []any,
map[string]any. Logical types use time.Time (UTC) for timestamps and
dates, time.Duration for time-of-day types, json.Number for decimals,
and avro.Duration for the duration logical type.
Encoding also accepts json.Number for any numeric type (supporting
json.Decoder.UseNumber() pipelines) and []byte for string fields (and
vice versa).
Struct fields are matched to Avro record fields by name. Use the avro struct
tag to control the mapping:
type Example struct {
Name string `avro:"name"` // maps to Avro field "name"
Ignored int `avro:"-"` // excluded from encoding/decoding
Inner Nested `avro:",inline"` // inline Nested's fields into this record
Value int `avro:"val,omitzero"` // encode zero value as Avro default
}The tag format is:
avro:"[name][,option][,option]..."
The name portion maps the struct field to the Avro field with that name. If
empty, the Go field name is used as-is. A tag of "-" excludes the field
entirely.
Supported options:
-
inline: flatten a nested struct's fields into the parent record, as if they were declared directly on the parent. The field must be a struct or pointer to struct. This works like anonymous (embedded) struct fields, but for named fields. When using inline, the name portion of the tag must be empty.
-
omitzero: when encoding, if the field is the zero value for its type (or implements an
IsZero() boolmethod that returns true), the Avro default value from the schema is used instead. This is useful for optional fields in["null", T]unions or fields with explicit defaults.
Embedded (anonymous) struct fields are automatically inlined — their fields are promoted into the parent as if declared directly. To prevent inlining an embedded struct, give it an explicit name tag:
type Parent struct {
Nested // inlined: Nested's fields are promoted
Other Aux `avro:"other"` // not inlined: treated as a single field
}When multiple fields at different depths resolve to the same Avro field name, the shallowest field wins. Among fields at the same depth, a tagged field wins over an untagged one.
SchemaFor infers an Avro schema from a Go struct type, using the same struct
tags as encoding/decoding:
type User struct {
Name string `avro:"name"`
Age int32 `avro:"age,default=18"`
Email *string `avro:"email"`
CreatedAt time.Time `avro:"created_at"`
}
schema := avro.MustSchemaFor[User](avro.WithNamespace("com.example"))This produces the equivalent of:
{
"type": "record",
"name": "User",
"namespace": "com.example",
"fields": [
{"name": "name", "type": "string"},
{"name": "age", "type": "int", "default": 18},
{"name": "email", "type": ["null", "string"]},
{"name": "created_at", "type": {"type": "long", "logicalType": "timestamp-millis"}}
]
}Go types map to Avro types automatically: *T becomes a ["null", T] union,
time.Time becomes timestamp-millis, and so on (see Type Mapping).
Additional tag options for schema inference:
| Tag | Example | Description |
|---|---|---|
default= |
avro:",default=0" |
Default value (must be last; scalars only) |
alias= |
avro:",alias=old" |
Field alias for schema evolution (repeatable) |
timestamp-micros |
avro:",timestamp-micros" |
Override logical type |
decimal(p,s) |
avro:",decimal(10,2)" |
Decimal logical type (required for *big.Rat) |
uuid |
avro:",uuid" |
UUID logical type |
date |
avro:",date" |
Date logical type |
Options:
WithNamespace(ns)sets the Avro namespace for the record.WithName(name)overrides the record name (defaults to the Go struct name).
Schema.Root() returns a SchemaNode representing the parsed schema. This
provides read access to all schema metadata including field types, logical
types, doc strings, and custom properties:
schema, _ := avro.Parse(schemaJSON)
root := schema.Root()
for _, f := range root.Fields {
fmt.Printf("field %s: type=%s\n", f.Name, f.Type.Type)
if cn, ok := f.Props["connect.name"].(string); ok {
fmt.Printf(" kafka connect type: %s\n", cn)
}
}SchemaNode can also be used to build schemas programmatically:
node := &avro.SchemaNode{
Type: "record",
Name: "User",
Fields: []avro.SchemaField{
{Name: "name", Type: avro.SchemaNode{Type: "string"}},
{Name: "age", Type: avro.SchemaNode{Type: "int"}, Default: 18},
},
}
schema, err := node.Schema()Logical types decode to their natural Go equivalents:
| Logical Type | Avro Type | Encode | Decode |
|---|---|---|---|
| date | int | time.Time, RFC 3339 or YYYY-MM-DD string, or int | time.Time (UTC) |
| time-millis | int | time.Duration or int | time.Duration |
| time-micros | long | time.Duration or int | time.Duration |
| timestamp-millis | long | time.Time, RFC 3339 string, or int | time.Time (UTC) |
| timestamp-micros | long | time.Time, RFC 3339 string, or int | time.Time (UTC) |
| timestamp-nanos | long | time.Time, RFC 3339 string, or int | time.Time (UTC) |
| local-timestamp-millis | long | time.Time, RFC 3339 string, or int | time.Time (UTC) |
| local-timestamp-micros | long | time.Time, RFC 3339 string, or int | time.Time (UTC) |
| local-timestamp-nanos | long | time.Time, RFC 3339 string, or int | time.Time (UTC) |
| uuid | string or fixed(16) | [16]byte or string | [16]byte (typed target) or string (any target) |
| decimal | bytes or fixed | *big.Rat, float64, numeric string, json.Number, or underlying type | *big.Rat, json.Number, or underlying type |
| duration | fixed(12) | avro.Duration or underlying type | avro.Duration or underlying type |
When encoding, timestamp and date fields accept RFC 3339 strings, and decimal fields accept float64 and numeric strings (e.g. "3.14"). Values that don't match the expected format fall through to the underlying type's encoder, which will return an error.
Unknown logical types are silently ignored per the Avro spec, and the underlying type is used as-is.
Avro data is always written with a specific schema — the writer schema. When you read that data later, your application may expect a different schema — the reader schema. You may have added a field, removed one, or widened a type from int to long.
Resolve bridges this gap. Given the writer and reader schemas, it returns a
new schema that decodes data in the old wire format and produces values in the
reader's layout:
- Fields in the reader but not the writer are filled from defaults.
- Fields in the writer but not the reader are skipped.
- Fields that exist in both are matched by name (or alias) and decoded, with type promotion applied where needed (e.g. int → long).
Suppose v1 of your application wrote User records with just a name:
var writerSchema = avro.MustParse(`{
"type": "record", "name": "User",
"fields": [
{"name": "name", "type": "string"}
]
}`)In v2 you added an email field with a default:
var readerSchema = avro.MustParse(`{
"type": "record", "name": "User",
"fields": [
{"name": "name", "type": "string"},
{"name": "email", "type": "string", "default": ""}
]
}`)
type User struct {
Name string `avro:"name"`
Email string `avro:"email"`
}To read old v1 data with your v2 struct, resolve the two schemas:
resolved, err := avro.Resolve(writerSchema, readerSchema)
var u User
_, err = resolved.Decode(v1Data, &u)
// u == User{Name: "Alice", Email: ""}The following type promotions are supported:
| Writer → Reader |
|---|
| int → long, float, double |
| long → float, double |
| float → double |
| string ↔ bytes |
CheckCompatibility checks whether two schemas are compatible without
building a resolved schema. The direction you check depends on the guarantee
you need:
// Backward: new schema can read old data.
avro.CheckCompatibility(oldSchema, newSchema)
// Forward: old schema can read new data.
avro.CheckCompatibility(newSchema, oldSchema)
// Full: check both directions.
avro.CheckCompatibility(oldSchema, newSchema)
avro.CheckCompatibility(newSchema, oldSchema)When working with a schema registry, schemas often reference types defined in
other schemas. SchemaCache accumulates named types across multiple Parse
calls so they can be resolved:
var cache avro.SchemaCache
// Parse referenced schema first — order matters.
_, err := cache.Parse(`{
"type": "record",
"name": "Address",
"fields": [{"name": "city", "type": "string"}]
}`)
// Now parse a schema that references Address.
schema, err := cache.Parse(`{
"type": "record",
"name": "User",
"fields": [
{"name": "name", "type": "string"},
{"name": "address", "type": "Address"}
]
}`)Parsing the same schema string multiple times returns the cached result,
handling diamond dependencies without caller-side deduplication. The returned
*Schema is independent of the cache and safe to use concurrently.
Register custom Go type conversions with NewCustomType for type-safe
primitive conversions, or CustomType for advanced cases:
type Money struct {
Cents int64
Currency string
}
moneyType := avro.NewCustomType[Money, int64]("money",
func(m Money, _ *avro.SchemaNode) (int64, error) { return m.Cents, nil },
func(c int64, _ *avro.SchemaNode) (Money, error) {
return Money{Cents: c, Currency: "USD"}, nil
},
)
schema := avro.MustParse(moneySchema, moneyType)
// Encode and decode — Money fields are automatically converted.
data, _ := schema.Encode(&order)
var out Order
schema.Decode(data, &out) // out.Price is Money{Cents: 500, ...}
// Works with SchemaFor too.
schema = avro.MustSchemaFor[Order](moneyType)A matching custom type replaces the built-in logical type deserializer. Decode callbacks receive raw Avro-native values (int64 for long, int32 for int, etc.). A nil Decode suppresses the built-in handler with zero overhead, producing raw values directly:
// Decode timestamps as raw int64 instead of time.Time.
schema := avro.MustParse(raw, avro.CustomType{
LogicalType: "timestamp-millis",
AvroType: "long",
})For property-based dispatch (e.g., Kafka Connect / Debezium types), use
an empty matching criteria with ErrSkipCustomType:
avro.CustomType{
Decode: func(v any, node *avro.SchemaNode) (any, error) {
name, _ := node.Props["connect.name"].(string)
switch name {
case "io.debezium.time.Timestamp":
return time.UnixMilli(v.(int64)).UTC(), nil
default:
return nil, avro.ErrSkipCustomType
}
},
}The ocf sub-package reads and writes Avro Object Container Files —
self-describing binary files that embed the schema in the header and store
data in compressed blocks.
var schema = avro.MustParse(`{
"type": "record",
"name": "User",
"fields": [
{"name": "name", "type": "string"},
{"name": "age", "type": "int"}
]
}`)
f, _ := os.Create("users.avro")
w, err := ocf.NewWriter(f, schema, ocf.WithCodec(ocf.SnappyCodec()))
if err != nil {
log.Fatal(err)
}
w.Encode(&User{Name: "Alice", Age: 30})
w.Encode(&User{Name: "Bob", Age: 25})
w.Close()
f.Close()f, _ := os.Open("users.avro")
r, err := ocf.NewReader(f)
if err != nil {
log.Fatal(err)
}
defer r.Close()
for {
var u User
err := r.Decode(&u)
if err == io.EOF {
break
}
if err != nil {
log.Fatal(err)
}
fmt.Println(u)
}The reader's Schema() method returns the schema parsed from the file header,
which you can pass as the writer schema to Resolve.
Built-in codecs: null (default, no compression), deflate
(DeflateCodec), snappy (SnappyCodec), and zstandard (ZstdCodec).
Custom codecs can be provided via the Codec interface.
NewAppendWriter opens an existing OCF for appending — it reads the header to
recover the schema, codec, and sync marker, then seeks to the end.
EncodeJSON is a schema-aware JSON serializer. By default it produces standard
JSON with bare union values and \uXXXX-encoded bytes:
// Standard JSON (default): bare unions
jsonBytes, err := schema.EncodeJSON(&user)
// {"name":"Alice","email":"a@b.com"}
// Avro JSON: unions wrapped as {"type_name": value}
jsonBytes, err = schema.EncodeJSON(&user, avro.TaggedUnions())
// {"name":"Alice","email":{"string":"a@b.com"}}DecodeJSON accepts both formats (tagged and bare unions) and all NaN/Infinity
conventions:
var user User
err = schema.DecodeJSON(jsonBytes, &user)Decode and DecodeJSON also accept TaggedUnions() to wrap union values
when decoding into *any:
var native any
schema.Decode(binary, &native, avro.TaggedUnions())
// native["email"] is map[string]any{"string": "a@b.com"}Encode and DecodeJSON accept both tagged and bare union input, so
tagged union output from Decode can round-trip through Encode directly.
Pass TagLogicalTypes() with TaggedUnions() to qualify union branch names
with their logical type (e.g. "long.timestamp-millis" instead of "long"),
matching the linkedin/goavro naming convention.
NaN and Infinity float values are encoded as "NaN", "Infinity", "-Infinity"
strings by default (Java Avro convention). Pass LinkedinFloats() for
the linkedin/goavro convention (null for NaN, ±1e999 for Infinity).
For sending self-describing values over the wire (as opposed to files, where OCF is preferred), use Single Object Encoding. Each message is a 2-byte magic header, an 8-byte CRC-64-AVRO fingerprint, and the Avro binary payload.
// Encode with fingerprint header
data, err := schema.AppendSingleObject(nil, &user)
// Decode (schema known)
_, err = schema.DecodeSingleObject(data, &user)
// Decode (schema unknown): extract fingerprint, look up schema
fp, payload, err := avro.SingleObjectFingerprint(data)
schema := registry.Lookup(fp) // your schema registry
_, err = schema.Decode(payload, &user)Canonical returns the Parsing Canonical Form
of a schema — a deterministic JSON representation stripped of doc, aliases,
defaults, and other non-essential attributes. Use it for schema comparison and
fingerprinting.
canonical := schema.Canonical() // []byte
// CRC-64-AVRO (Rabin) — the Avro-standard fingerprint
fp := schema.Fingerprint(avro.NewRabin())
// SHA-256 — common for cross-language registries
fp256 := schema.Fingerprint(sha256.New())Encode and decode errors can be inspected with errors.As:
*SemanticError: type mismatch between Go and Avro (includes a dotted field path for nested records, e.g."address.zip").*ShortBufferError: input truncated mid-value.*CompatibilityError: schema evolution incompatibility (fromResolveorCheckCompatibility).
Struct field access uses unsafe pointer arithmetic (similar to
encoding/json v2) to avoid reflect.Value overhead on every encode/decode.
All schemas, type mappings, and codec state are cached after first use so
repeated operations pay no extra allocation cost.