Accelerate the development of Generative AI application with GitHub Models

  • Thread starter Thread starter kinfey
  • Start date Start date
K

kinfey

Screenshot 2024-08-02 at 5.29.41 PM.png



The first step in developing generative AI applications is to choose a model. How to choose a model is the key. This includes

  1. When we combine application development with business scenarios, there are many comparisons, such as the generation effects of the same prompt words under different models.
  2. Quick comparison and switching of multiple models
  3. How different models adapt to new application frameworks and solutions to complete projects more effectively.

The release of GitHub Models plays a very important role for developers and different development teams to more effectively select models in the process of developing applications and create applications based on different application frameworks. Let’s take a look at how I use GitHub Models to complete development in different scenarios.

Model comparison


In GitHub Models, through the provided playground, we can complete the comparison of the same prompt for different models.

Let’s take a look at the comparison between Phi-3-mini and Mistral Nemo
01.png

Judging from the results, this is an evenly matched result.

Quick comparison and switching of multiple models


Above, we switched models in the playground to compare different models under the same prompt. For development, a more direct approach may be required. With the Azure AI Inference SDK you can quickly switch to different models. You can choose Python, JavaScript, and REST access methods by selecting Code.

If we choose the Phi-3-mini scenario, we can choose to obtain the access method in Code

02.png

Of course, you can directly and seamlessly access the programming environment through Codespace.

Adaptation to different application frameworks


Generative AI has different application frameworks combined with models to complete applications, such as GraphRAG. We can use the REST interface provided by GitHub Models to test model solutions other than GPT-4o, such as selecting the latest Meta LLama 3.1 405b Instruct. If the local deployment of this model has been limited by computing power, it will be difficult for individuals and small teams to adopt it. But based on the interface provided by GitHub Models, we can complete the test in the local environment very simply

  1. Configure the environment

Install the GraphRAG Python library

Code:
pip install graphrag -U
  1. Create a GraphRAG project

Code:
mkdir -p ./ragmd/input

python -m graphrag.index --init --root ./ragmd
  1. Modify settings.yaml

Code:
encoding_model: cl100k_base
skip_workflows: []
llm:
 api_key: ${GRAPHRAG_API_KEY}
 type: openai_chat # or azure_openai_chat
 model: meta-llama-3.1-405b-instruct
 model_supports_json: true # recommended if this is available for your model.
 max_tokens: 4000
 api_base: https://models.inference.ai.azure.com

parallelization:
 Stagger: 0.3

async_mode: threaded # or asyncio

embeddings:
 async_mode: threaded # or asyncio
 llm:
 api_key: ${GRAPHRAG_API_KEY}
 type: openai_embedding # or azure_openai_embedding
 model: jinaai
 api_base: http://localhost:5146/v1

Note Please configure GitHub Tokens in .env

  1. Run

Code:
python -m graphrag.index --root ./ragmd

Test Results

Code:
python -m graphrag.query --root ./ragmd --method global "What's GraphRAG"

03.png

Through GitHub Models, we can quickly use the provided models for model comparison and application development environment testing, which allows model and application testing to be completed more efficiently and quickly in environments with limited computing power.

Learning Resources


  1. Sign Up Build software better, together


  2. Introducing GitHub Models: A new generation of AI engineers building on GitHub Introducing GitHub Models: A new generation of AI engineers building on GitHub


  3. Understand Phi-3 GitHub - microsoft/Phi-3CookBook: This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.


  4. Learn about GraphRAG Microsoft on GitHub

Continue reading...
 
Back
Top