Files
2024-09-22 04:17:12 -04:00
..
2024-09-22 04:17:12 -04:00

Code

# %%

import os; os.system('pip install -q requests bs4 lxml pandas')
import requests, bs4, json, pandas as pd

def ollama(prompt):
    return requests.post(
        "https://ollama.yauk.tv/api/generate",
        json={
            "model": "llama3.1",
            "format": "json",
            "prompt": prompt,
            "stream": False,
            "options": {
                "temperature": 0,
                "num_ctx": 32768
            }
        }
    ).json()['response']

# %%

url = 'https://ollama.com/library'
response = requests.get(url)
soup = bs4.BeautifulSoup(response.text, 'lxml')
tags = ''.join(str(tag.prettify()) for tag in soup.find_all('li', class_='flex')[:20])
print(tags)

# %%

result = ollama(f''' Parse the following HTML snippet and extract the information into a JSON format. Output only the JSON data, without any additional text, explanation, or formatting.

HTML to analyze: {tags} ''')

df = pd.DataFrame(list(json.loads(result).values())[0])
df

Output

name description size pulls tags updated
0 llama3.1 Llama 3.1 is a new state-of-the-art model from Meta available in 8B, 70B and 405B parameter sizes. ['8B', '70B', '405B'] 5.3M 94 6 days ago
1 gemma2 Google Gemma 2 is a high-performing and efficient model available in three sizes: 2B, 9B, and 27B. ['2B', '9B', '27B'] 1.2M 94 6 days ago
2 qwen2.5 Qwen2.5 models are pretrained on Alibaba's latest large-scale dataset, encompassing up to 18 trillion tokens. ['0.5B', '1.5B', '3B', '7B', '14B', '32B', '72B'] 289.1K 133 3 days ago
3 phi3.5 A lightweight AI model with 3.8 billion parameters with performance overtaking similarly and larger sized models. ['3B'] 53.9K 17 4 weeks ago
4 nemotron-mini A commercial-friendly small language model by NVIDIA optimized for roleplay, RAG QA, and function calling. [] 7,650 17 2 days ago
5 mistral-small Mistral Small is a lightweight model designed for cost-effective use in tasks like translation and summarization. ['22B'] 5,816 17 4 days ago
6 mistral-nemo A state-of-the-art 12B model with 128k context length, built by Mistral AI in collaboration with NVIDIA. ['12B'] 202.2K 17 4 hours ago
7 deepseek-coder-v2 An open-source Mixture-of-Experts code language model that achieves performance comparable to GPT4-Turbo in code-specific tasks. ['16B', '236B'] 294.4K 65 3 months ago
8 mistral The 7B model released by Mistral AI, updated to version 0.3. ['7B'] 3.4M 84 4 months ago
9 mixtral A set of Mixture of Experts (MoE) model with open weights by Mistral AI in 8x7b and 8x22b parameter sizes. ['8x7B', '8x22B'] 415.8K 69 5 months ago
10 codegemma CodeGemma is a collection of powerful, lightweight models that can perform a variety of coding tasks like fill-in-the-middle code completion, code generation, natural language understanding, mathematical reasoning, and instruction following. ['2B', '7B'] 282.5K 85 5 months ago
11 command-r Command R is a Large Language Model optimized for conversational interaction and long context tasks. ['35B'] 203.5K 32 3 weeks ago
12 command-r-plus Command R+ is a powerful, scalable large language model purpose-built to excel at real-world enterprise use cases. ['104B'] 95.8K 21 3 weeks ago
13 llava 🌋 LLaVA is a novel end-to-end trained large multimodal model that combines a vision encoder and Vicuna for general-purpose visual and language understanding. ['7B', '13B', '34B'] 1.1M 98 7 months ago
14 llama3 Meta Llama 3: The most capable openly available LLM to date. ['8B', '70B'] 6.2M 68 4 months ago
15 gemma Gemma is a family of lightweight, state-of-the-art open models built by Google DeepMind. ['2B', '7B'] 4.1M 102 5 months ago
16 qwen Qwen 1.5 is a series of large language models by Alibaba Cloud spanning from 0.5B to 110B parameters. ['0.5B', '1.8B', '4B', '32B', '72B', '110B'] 4M 379 3 months ago
17 qwen2 Qwen2 is a new series of large language models from Alibaba group. ['0.5B', '1.5B', '7B', '72B'] 3.7M 97 3 months ago
18 llama2 Llama 2 is a collection of foundation language models ranging from 7B to 70B parameters. ['7B', '13B', '70B'] 2.1M 102 7 months ago