Stable Release - v4.5.0 #31
MeridianAlgo-Developer
started this conversation in
General
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Meridian.AI: Financial Intelligence MoE Model
Meridian.AI is a high-performance financial language model utilizing a Grouped Query Attention (GQA) backed Sparse Mixture-of-Experts (SMoE) architecture. Based on the OpenMoE-650M foundation, the model is specifically engineered for financial intelligence, high-precision quantitative reasoning, algorithmic math tasks, and capital markets analytics.
It introduces an innovative training paradigm optimized for continuous, hourly execution on standard CPU runners, enabled through the application of Elastic Weight Consolidation (EWC) to prevent catastrophic forgetting.
Model Access & Hugging Face Deployment
The latest live training checkpoints and full model weights are perpetually synchronized and stored on the Hugging Face Hub.
Hugging Face Repository: https://huggingface.co/MeridianAlgo/MeridianAI
You can download the model instantly using the
huggingface_hubPython toolkit or directly through standardtransformerspipelines.Architecture and Technical Foundations
Meridian.AI leverages a custom Sparse Mixture-of-Experts architecture to maximize knowledge capacity while maintaining extreme efficiency during inference and training loops.
1. Sparse Mixture-of-Experts (SMoE)
The model functions on a highly sparse active parameter plane. By utilizing a sparse gateway system with distinct experts and activating only a small subset of parameters per token, the model achieves the representational capacity of a much larger dense network without the associated computational cost. This makes it ideal for rapid deployment on standard CPU environments.
2. Elastic Weight Consolidation (EWC)
To support perpetual hourly learning natively inside GitHub Actions, the model utilizes EWC (Elastic Weight Consolidation). This technique calculates the Fisher Information Matrix to identify weights critical to previously learned financial knowledge. During incremental training, an active penalty restricts changes in these vital weights, ensuring the model retains its core financial reasoning capabilities while safely adapting to new real-time market data.
3. Quantitative Reasoning & Numeracy Encoders
Unlike generic general-purpose models, Meridian.AI boasts embedded numeracy encoders mapping magnitude signals directly into the dense representation. It is continuously fine-tuned on a specialized curriculum of financial instruction sets, real-time news data, and mathematical reasoning arrays to ensure immense precision when handling quantitative logic and complex financial analysis.
Model Specifications
Automated Lifecycle & Deployment
The repository possesses a fully autonomous lifecycle governed by GitHub Actions:
safetensorsto the Hub post-training.Getting Started
Environment Setup
# Clone the repository and install the custom dependencies python -m pip install -r requirements.txtContinual Training Loop
The
train.pyscript manages the primary continual learning loop. Set your environments and execute the automated hourly runner format:License
This project is completely open source and distributed under the MIT License.
made with love by meridianalgo 🩵
This discussion was created from the release Stable Release - v4.5.0.
Beta Was this translation helpful? Give feedback.
All reactions