Skip to content
View Mmorgan-ML's full-sized avatar

Block or report Mmorgan-ML

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Mmorgan-ML/README.md

Hi, I'm Michael Christian Morgan.

Independent Machine Learning Researcher

📍 South Korea (Open to Remote Roles)

I am a researcher bridging the gap between Linguistics, Complex Systems Biology, and Artificial Intelligence.

With an academic foundation in East Asian Linguistics and Articulatory Phonetics from the University of Georgia, I approach Large Language Models not just as statistical predictors, but as dynamic systems akin to biological evolution and physical processes. My work focuses on escaping local minima in architectural design by applying principles from nature, physics, and "weird math" to deep learning.


🔬 Research Focus & Technical Stack

I specialize in Python and PyTorch. My work often requires going deeper than standard libraries, including writing custom C++/CUDA kernels for optimizing novel architectures like my Neuromodulatory Control Networks.

My current research interests include:

  • Physics-Inspired Architectures: Energy-based models, Hamiltonian dynamics, and thermodynamic approaches to loss landscapes.
  • Biologically Plausible AI: Neuromodulation, hypernetworks, and synaptic plasticity simulations.
  • Evolutionary Computing: Genetic algorithms applied to Transformer hyperparameter optimization and Artificial Life simulations.
  • Linguistics: Viewing tokenization through the lens of articulatory phonetics and continuous dynamic systems.

⚡ Featured Projects

A massive-scale evolutionary simulator for hyperparameter optimization and artificial life.

evolution_lab image

This project is a comprehensive framework that pits populations of "MicroTransformers" (10k parameters) against each other in a survival-of-the-fittest environment based on validation loss performance.

  • The Genome: Agents possess 17 "genes" representing initialization hyperparameters.
  • Lifecycle: Agents fight, breed, mutate, clone, and evolve over generations.
  • Gene Forge: A tool for manual gene editing and Horizontal Gene Transfer (stealing genes from other successful agents).
  • Stress Chamber: A rigorous testing module for collecting performance data on specific genomes across random seeds.
  • Speciation Dashboard: Uses PCA to map 17-dimensional gene data onto a 3D interactive graph. It utilizes clustering to automatically categorize agents into taxonomic ranks (Kingdom, Phylum, Class, etc.), visualizing how distinct survival strategies evolve divergently.

A novel LLM architecture inspired by the human brain's neuromodulatory systems.

Successfully trained 18M param model on 1 epoch of TinyStories, achieving 4.5 PPL.

Neuromodulation (Fig 1) Token-Level Neuromodulatory Dynamics. Heatmaps display the NCN output values for Layer Gain (Top), Attention Precision (Middle), and FFN Gating (Bottom). X-axis represents the token sequence; Y-axis represents Layer Depth (0-5). perplexity_graph (Fig 2) Training Convergence Analysis. The NCN model demonstrates rapid perplexity reduction, stabilizing significantly below standard baselines for this parameter class. The dark blue line represents the macro-trend (span=118), showing a smooth descent without the volatility typically associated with hypernetwork training. analysis_grammar_95 (Fig 3) Grammar/Syntax Convergence (95%). The dashed blue line indicates a steep power-law fit, suggesting rapid acquisition of surface-level statistics. analysis_intellectual_99 (Fig 4) Intellectual Convergence (99%). The "Active Phase" (Red) tightly hugs the power law, indicating sustained learning of higher-order logic without premature plateaus.

This architecture moves beyond standard Transformers by implementing a global modulation mechanism.

  • Mechanism: Functions similarly to a hypernetwork but outputs modulation signals that dynamically adjust the Temperature, Gain, and FFN Gating of the main network blocks.
  • Implicit Learning: The network learns to contextually identify which modulation signals minimize loss without explicit supervision.
  • Optimization: Includes custom C++/CUDA kernels to handle the specific computational requirements of the modulation layers.
  • (Paper currently in pre-print phase, includes preliminary empirical results).

Available on PyPI.

image

A novel inference sampler designed to prevent LLM looping and degradation.

  • Method: Monitors the entropy of the model's output in real-time.
  • Intervention: Injects non-destructive Gaussian noise directly into the Key-Value (KV) cache when entropy thresholds suggest loop formation, effectively "kicking" the model out of repetitive states without breaking coherence.

📜 Background & Experience

  • Linguistics & Phonetics: My background gives me a unique perspective on NLP. I view language processing as a constraint system rooted in physical production (articulatory phonetics) rather than purely symbolic manipulation.
  • Patent Translation: Extensive experience translating technical patents (English/Korean/Japanese), requiring extreme attention to detail and technical literacy.
  • Global Business: Experience in Duty Free Trade, connecting international suppliers to the Korean market.

📫 Connect

I am currently open to job offers and research collaborations (Remote preferred).

Pinned Loading

  1. Neuromodulatory-Control-Networks Neuromodulatory-Control-Networks Public

    Neuromodulatory Control Networks (NCNs), a novel LLM architectural modification inspired by the neuromodulatory systems in the vertebrate brain.

    Python 21 4

  2. Neural-Speciation Neural-Speciation Public

    An evolutionary engine for micro-transformers that optimizes initialization hyperparameters. Uses unsupervised clustering to discover distinct "species" of learning strategies that convergently evo…

    2

  3. Phase-Slip-Sampler Phase-Slip-Sampler Public

    An entropy-aware sampling strategy that injects nondestructive noise into the KV Cache to break repetition loops and force semantic state transitions in LLMs.

    Python 1