About Me

  • Full Name:Hiva Mohammadzadeh
  • Email:hiva@berkeley.edu
  • Website:hivam.org

Hello There! Welcome to my personal webpage!

Please let me take this opportunity to tell you about myself and my past experiences.

My childhood love for Mathematics and Computers drove me from Tehran, Iran to California, USA where I am passionate about making the world a better place through Natural Language Processing, Machine Learning, and Software Engineering. In my past research and interning experiences, I have had the chance to apply the knowledge that I have gained in my vigorous classes to real life problems and projects.

I am motivated to continuously expand my knowledge and grow in my profession, leveraging my fast-learning abilities, organizational aptitude, leadership qualities, and technical expertise.

I am a recent graduate of EECS at UC Berkeley and currently a Machine Learning Researcher at UC Berkeley's Artificial Intelligence Lab advised by Kurt Keutzer.

I am currently working on Efficient Deep Learning with a focus on Natural Language Processing and Large Language Models. In particular, I am interested in sparsity, quantization, and new training/fine-tuning methods to enable models that can learn more efficiently. I am also developing algorithms to compress large neural network models, focusing on reducing inference time and improving training efficiency.

I have also been exploring AI Agents, focusing on the key components required to build them and the system level decisions that effect their performance (Paper Coming December 2024)

My Resume

  • Education

  • Bachelor of Science: Electrical Engineering and Computer Sciences

    University of California, Berkeley - August 2021 - December 2023
  • Associate Degree for Transfer in Computer Science and Programming

    Los Angeles Pierce Collegee - August 2019 - June 2021
  • Publications

  • Squeezed Attention: Accelerating Long Context Length LLM Inference

    Coleman Hooper*, Sehoon Kim*, Hiva Mohammadzadeh, Monishwaran Maheswaran, June Paik, Michael W. Mahoney, Kurt Keutzer, Amir Gholami

    Submitted to MLSys 2025 and on Arxiv.

  • KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization

    Coleman Hooper, Sehoon Kim, Hiva Mohammadzadeh, Michael W. Mahoney, Yakun Sophia Shao, Kurt Keutzer, Amir Gholami

    Accepted as a poster to NeurIPS 2024 and on Arxiv.

  • SPEED: Speculative Pipelined Execution for Efficient Decoding

    Coleman Hooper, Sehoon Kim, Hiva Mohammadzadeh, Hasan Genc, Kurt Keutzer, Amir Gholami, Sophia Shao

    Accepted as a poster presentation to ENLSP NeurIPS 2023 Workshop and on Arxiv.

  • Plume-induced delamination initiated at rift zones on Venus

    Andrea Adams, Dave Stegman, Hiva Mohammadzadeh, Suzanne Smrekar, Paul Tackley

    Published to the Journal of Geophysical Research: Planets.

  • Presentations and Tutorials

  • NLP Presentatios

    Spring 2023 - Summer 2023

    This Repository includes survey of all the Large Language Models, Benchmarks, Normalization techniques, and Activation Functions.

  • Supervised Independent Study with Professor Joseph Gonzalez

    Spring 2023

    This Repository includes a survey of all key papers in NLP.

  • Research Experience

  • Machine Leaning Researcher

    Pallas Group at UC Berkeley AI Research Lab (BAIR) and SLICE Lab - February 2023 - Present

    Mentor: Professor Kurt Keutzer.

    Building efficient LLM-based systems and working on a survey of AI Agents as the first author

    Collaborated on Squeezed Attention, a technique to accelerate LLM inference in applications where a large portion of the input prompt is fixed. (Submitted to MLSys 2025)

    Contributed as co-author to KVQuant which enables large context length inference and allows for serving LLaMA-7B with 1M tokens on a single A100!

    Built an architecture to accelerate generative LLM inference by 40% as co-author for a published paper.

    Innovated new approaches for efficient deep learning and NLP.

  • Undergraduate NLP Researcher

    Sky Computing Lab at UC Berkeley - August 2022 - February 2023

    Mentor: Professor Joseph Gonzalez (UC Berkeley)

    Completed individual course of study with Prof. Joseph Gonzalez to design project for efficient language models. Fine and prompt tuned language models to build two scientific article-focused chatbots for students and researchers.

  • Undergraduate Researcher

    Computational Infrastructure for Geodynamics, NSF, UCSD, NASA/JPL - June 2021 - October 2021

    Mentors: Prof. Dave Stegman (UCSD) and Dr. Sue Smrekar (NASA, JPL)

    Built and analyzed a model of Venus on supercomputers using Python and Fortran with Prof. Dave Stegman. Found that plume-assisted tectonic subduction happens 80% faster than hypothesized while advised by Dr. Sue Smrekar.

    Co-authored scientific paper in support of NASA’s Venus VERITAS mission of NASA/JPL.

  • Work Experience

  • Research And Development Intern - Data Science and Power Systems Modeling Engineer

    Span.io (Series B Startup) - May 2016 - September 2022

    Designed and implemented python software to solve Nonlinear Differential Equations to speed up analytics by 75%. Simulated home appliance power consumption using the Span Panel data to inform next product iteration. Analyzed the value of electrification technologies with processing usage data of the Panel in Python and Snowflake to predict the best product sales with the goal to generate higher revenue for the company.


Selected Skills

Python

Java

Software Development

Research

C/C++

SQL

Contact Me!

Please do not hesitate to contact me via email or linkedin.

Contact Details