Meta introduces new byte-based language model architecture
Meta and the University of Washington have developed a new AI architecture called Byte latent transformer (BLT) that processes language without traditional tokenization. As reported by Ben Dickson for VentureBeat, BLT works directly with raw bytes instead of predefined tokens, making it more versatile and efficient. The system uses three transformer blocks: two lightweight encoder/decoder … Read more