Llama

Meta Llama 3, the next generation of state-of-the-art open source large language model. This release features pretrained and instruction-fine-tuned language models with 8B and 70B parameters that can support a broad range of use cases. 

Llama 3 uses a tokenizer with a vocabulary of 128K tokens that encodes language much more efficiently, which leads to substantially improved model performance. 

We are enthusiasts and lovers of this incredible tool. On this website we will be exploring its possibilities and sharing latest news about Llama 3 Model.


  • Optimizing RAG with LLamaIndex

    A cool trick you can use to improve retrieval performance in your RAG pipelines is fine-tune the embedding model (bi-encoder) based on labels from a cross-encoder 💡 Cross-encoders are crucial for reranking but are way too slow for retrieving over large numbers of documents. This fine-tuning technique gives you all the speed advantages of direct…

    Continue reading

  • Run LLama 3 on iPhone 15 Pro

    In addition to other improvements, current release enables running Meta Llama 2 7B efficiently on devices like the iPhone 15 Pro, Samsung Galaxy S24 and other edge devices — it also includes early support for Llama 3 8B. More details on ExecuTorch Alpha below. ExecuTorch ExecuTorch Alpha is focused on deploying large language models and…

    Continue reading

  • LLama 3 vs GPT-4

    Llama 3 is a cutting-edge large language model introduced by Meta AI on April 18, 2024. This model family offers three sizes: 8B, 70B, and 400B. The 8B and 70B versions are available to the public, whereas the 400B version is currently undergoing training. Llama 3 boasts benchmark scores that match or surpass those of…

    Continue reading