Close Menu
    Facebook X (Twitter) Instagram
    • About Us
    • Privacy Policy
    • Write For Us
    • Cookie Policy
    • Terms and Conditions
    Facebook X (Twitter) Instagram
    CompaniesHistory.com – The largest companies and brands in the world
    • Who Owns
    • AI
    • Business
      • Finance
    • Technology
      • Crypto
      • Software
      • Biotech
    • iGaming
    • Others
      • Real Estate
      • FMCG
      • Logistics
      • Lifestyle
    • Blog
    • Contact Us
    CompaniesHistory.com – The largest companies and brands in the world
    Home»AI»Galactica Net Worth, Revenue, Marketcap, Competitors 2026

    Galactica Net Worth, Revenue, Marketcap, Competitors 2026

    DariusBy DariusDecember 10, 2025No Comments5 Mins Read
    Meta Galactica AI model 120 billion parameters trained on 48 million scientific papers and 106 billion tokens, public demo removed after 3 days November 2022 due to accuracy concerns, outperformed GPT-3 on LaTeX equations 68.2% vs 49.0%, achieved 77.6% PubMedQA and 52.9% MedMCQA accuracy, six model variants 125 million to 120 billion parameters, trained on 360 million citations 50 million references 8 million lecture notes, required 128 NVIDIA A100 80GB nodes for training, Meta AI investment $39.2 billion 2024 projected $66-72 billion 2025, models remain available on GitHub and HuggingFace, Meta AI serves 700 million monthly active users.
    Share
    Facebook Twitter LinkedIn Pinterest Email

    Meta’s Galactica reached 120 billion parameters and trained on 48 million scientific papers before its public demo was pulled after just three days in November 2022. The model outperformed GPT-3 on LaTeX equation tasks by 19.2 percentage points and set new records on medical question-answering benchmarks. This article examines the verified statistics behind Galactica’s architecture, performance, and lasting influence on Meta’s AI strategy.

    Galactica Key Statistics

    • Galactica trained on 48 million+ scientific papers and 106 billion tokens as of 2022
    • The largest Galactica model contains 120 billion parameters
    • Galactica achieved 77.6% accuracy on PubMedQA, setting a new state-of-the-art record
    • Meta removed the public demo after 3 days following accuracy concerns
    • Meta’s AI capital expenditure reached $39.2 billion in 2024, partly influenced by Galactica lessons

    Galactica Training Data and Corpus

    The training corpus, internally called “NatureBook,” contained over 48 million scientific papers, 360 million in-context citations, and 50 million unique references. Meta’s researchers curated this dataset specifically for scientific knowledge tasks rather than relying on web-crawled data.

    The specialized corpus included 2 million code samples and 8 million lecture notes and textbooks. Training ran for approximately 450 billion tokens across 4.25 epochs without overfitting issues common to generic language models.

    Galactica Model Architecture Statistics

    Meta released Galactica as a family of six models ranging from 125 million to 120 billion parameters. The smallest variant, Galactica Mini, targeted lightweight applications while the 120 billion parameter flagship handled complex reasoning tasks.

    Model Variant Parameters Primary Use Case
    Galactica Mini 125 million Lightweight applications
    Galactica Base 1.3 billion Standard research tasks
    Galactica Standard 6.7 billion Enhanced capabilities
    Galactica Large 30 billion Complex reasoning
    Galactica Huge 120 billion Maximum performance

    Training the 120 billion parameter model required 128 NVIDIA A100 80GB nodes. Meta optimized inference to run on a single A100 node, prioritizing accessibility for researchers. Performance scaled consistently with model size across all benchmark tasks.

    Galactica Benchmark Performance

    Galactica outperformed GPT-3 on LaTeX equation generation with a score of 68.2% compared to 49.0%. On the MATH benchmark, it achieved 20.4% accuracy versus PaLM 540B’s 8.8%, representing an 11.6 percentage point advantage.

    The model set new state-of-the-art results on medical benchmarks including 77.6% on PubMedQA and 52.9% on MedMCQA Dev. Mathematical MMLU scores reached 41.3%, exceeding DeepMind’s Chinchilla by 5.6 percentage points.

    Citation Prediction Accuracy

    Citation prediction accuracy ranged from 36.6% to 69.1% depending on the evaluation dataset. Researchers noted a documented bias toward highly-cited papers over lesser-known work in the original technical paper.

    Galactica Timeline and Public Response

    Meta launched the Galactica public demo on November 15, 2022. Within hours, researchers documented instances of the model generating factually incorrect scientific content. Meta removed the demo on November 17, 2022, after just three days of public availability.

    Michael Black, Director of the Max Planck Institute for Intelligent Systems, warned the model could enable “deep scientific fakes.” Meta’s VP of AI Research Joelle Pineau later acknowledged the company “probably misjudged” public expectations for scientific AI.

    ChatGPT launched 15 days after Galactica’s removal, followed by Meta’s Llama release approximately 100 days later. The Galactica experience directly informed Meta’s more cautious approach to subsequent model releases.

    Galactica Infrastructure Requirements

    The 120 billion parameter model required 128 NVIDIA A100 nodes with 80GB memory each for training. Meta deliberately constrained maximum model size to ensure single-node inference capability for broader research accessibility.

    Specification Requirement
    Training Hardware (120B) 128 NVIDIA A100 80GB nodes
    Inference Hardware (120B) Single A100 node
    Training Duration 450 billion tokens
    Memory Per Node 80 GB maximum

    The metaseq library developed by Meta AI’s NextSys team powered training infrastructure. This same foundation later supported Llama model development.

    Meta AI Investment After Galactica

    Meta reported $39.2 billion in AI capital expenditures for 2024, with Q4 alone reaching $14.8 billion. The company projects spending between $66 billion and $72 billion in 2025, representing approximately 70% year-over-year growth.

    CEO Mark Zuckerberg disclosed Meta expects to spend at least $600 billion on U.S. data centers and AI infrastructure by 2028. Meta AI now serves over 700 million monthly active users with projections suggesting 1 billion users within 2025.

    Galactica Legacy and Open Source Availability

    All six Galactica model variants remain available on GitHub and HuggingFace. Meta’s VP of AI Research confirmed ongoing requests from researchers seeking access to the models years after the original release.

    The Llama 3.1 model has recorded 5.8 million HuggingFace downloads, demonstrating the success of Meta’s revised release strategy informed by Galactica’s reception. Meta’s generative AI tools now attract more than 4 million advertisers.

    FAQ

    How many parameters does Galactica have?

    Galactica ranges from 125 million to 120 billion parameters across six model variants. The largest 120 billion parameter version delivers maximum performance for complex scientific reasoning tasks.

    Why was Galactica removed after three days?

    Meta removed Galactica after researchers documented instances of factually incorrect scientific content and fictitious citations. The gap between public expectations and research reality proved too significant.

    What data was Galactica trained on?

    Galactica trained on 48 million scientific papers, 106 billion tokens, 360 million citations, 2 million code samples, and 8 million lecture notes curated in a corpus called NatureBook.

    How does Galactica compare to GPT-3?

    Galactica outperformed GPT-3 on LaTeX equation tasks by 19.2 percentage points, scoring 68.2% compared to GPT-3’s 49.0%. It also set new records on medical question-answering benchmarks.

    Is Galactica still available to use?

    All six Galactica model variants remain available as open-source downloads on GitHub and HuggingFace. The public demo was removed, but researchers can still access the model weights.

    Sources

    • arXiv Galactica Technical Paper
    • VentureBeat Meta AI Coverage
    • AI Business Galactica Report
    • TechCrunch Meta Investment Analysis
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Darius
    • Website
    • Facebook
    • X (Twitter)
    • Instagram
    • LinkedIn

    I've spent over a decade researching and documenting the stories behind the world's most influential companies. What started as a personal fascination with how businesses evolve from small startups to global giants turned into CompaniesHistory.com—a platform dedicated to making corporate history accessible to everyone.

    Related Posts

    Midjourney Statistics And User Demographics 2026

    January 24, 2026

    Florence Statistics 2026

    January 23, 2026

    Adobe Firefly Statistics And User Trends 2026

    January 22, 2026

    AlphaCode Statistics 2026

    January 21, 2026
    CompaniesHistory.com – The largest companies and brands in the world
    Facebook X (Twitter) Instagram YouTube LinkedIn
    • About Us
    • Privacy Policy
    • Write For Us
    • Cookie Policy
    • Terms and Conditions

    Type above and press Enter to search. Press Esc to cancel.