Leanpub Header

Skip to main content

Computational Neuroscience

The Science of Thought From Synapses to Supercomputers

A Comprehensive Textbook on the Mathematical and Computational Foundations of Brain Function

The brain is not just an organ—it is a computational system, processing vast amounts of information through dynamic neural networks. How does the brain encode, store, and retrieve information? How do electrical and chemical signals translate into thought, perception, and decision-making?

Minimum price

$19.00

$29.00

You pay

$29.00

Author earns

$23.20
$

...Or Buy With Credits!

You can get credits with a paid monthly or annual Reader Membership, or you can buy them here.
PDF
About

About

About the Book

A Comprehensive Textbook on the Mathematical and Computational Foundations of Brain Function

The brain is not just an organ—it is a computational system, processing vast amounts of information through dynamic neural networks. How does the brain encode, store, and retrieve information? How do electrical and chemical signals translate into thought, perception, and decision-making?

Computational neuroscience bridges the gap between biology and mathematics, offering a quantitative framework to study neural activity, learning, and cognition. This textbook provides an in-depth exploration of the mathematical, theoretical, and algorithmic principles governing the brain's computational processes.

With a focus on rigorous analysis and real-world applications, Computational Neuroscience: The Science of Thought From Synapses to Supercomputers equips students, researchers, and professionals with the essential knowledge to understand and model brain function.

Inside This Book
Foundations of Neural Computation

  • The Neuron as a Computational Unit – Understand the biophysical principles of synaptic transmission and action potentials.
  • Mathematical Models of Neurons – Explore Hodgkin-Huxley equations, integrate-and-fire models, and stochastic neuron models.
  • Information Theory & Neural Encoding – Learn how the brain processes and represents sensory information.

Neural Networks & Large-Scale Brain Computation

  • Network Dynamics & Functional Connectivity – Discover how neurons interact to form large-scale networks.
  • Hebbian Learning & Synaptic Plasticity – Examine how experience and adaptation shape neural circuits.
  • Bayesian Inference in Neural Computation – Apply probabilistic models to decision-making and perception.
  • Spiking Neural Networks & Temporal Coding – Analyze the role of precise timing in neural information processing.

Modeling Cognitive Functions

  • Neural Mechanisms of Perception – Understand how the brain processes vision, sound, and multisensory integration.
  • Computational Models of Memory – Study working memory, long-term memory storage, and recall mechanisms.
  • Decision-Making & Reinforcement Learning – Explore how neural circuits optimize choices and actions.

Artificial Intelligence & Brain-Inspired Computing

  • Deep Learning & Neural Networks – Examine how biological computation has influenced AI and machine learning.
  • Neuromorphic Engineering & Brain-Computer Interfaces – Discover how brain-inspired hardware is shaping the future of computing.
  • Emerging Trends in Neurotechnology – Investigate advancements in brain simulation, cognitive augmentation, and neuroprosthetics.

Who This Book Is For

📘 Students & Researchers – A structured, research-backed introduction to the field of computational neuroscience.

📘 Neuroscientists & Cognitive Scientists – A rigorous approach to mathematical modeling and neural simulation.

📘 AI & Machine Learning Professionals – A deep dive into biologically inspired computing and its applications.

📘 Physicists, Mathematicians & Engineers – A quantitative approach to understanding the complex dynamics of brain computation.

Get your copy today and explore the science behind intelligence!

Share this book

Categories

Author

About the Author

gareth thomas

Gareth Morgan Thomas is a qualified expert with extensive expertise across multiple STEM fields. Holding six university diplomas in electronics, software development, web development, and project management, along with qualifications in computer networking, CAD, diesel engineering, well drilling, and welding, he has built a robust foundation of technical knowledge.

Educated in Auckland, New Zealand, Gareth Morgan Thomas also spent three years serving in the New Zealand Army, where he honed his discipline and problem-solving skills. With years of technical training, Gareth Morgan Thomas is now dedicated to sharing his deep understanding of science, technology, engineering, and mathematics through a series of specialized books aimed at both beginners and advanced learners.

Contents

Table of Contents

Chapter 1. Introduction to Computational Neuroscience

Section 1. Scope and Significance of the Field

  • Definition and importance of computational neuroscience
  • The role of modeling in understanding brain function
  • Applications in AI, medicine, and neuroengineering

Section 2. Interdisciplinary Nature and Historical Overview

  • Relationship with neuroscience, physics, mathematics, and AI
  • Influence of cognitive science and philosophy
  • Key research milestones in the field

Chapter 2. Historical Perspectives and the Evolution of Neural Modeling

Section 1. Milestones in Neuroscience and Computational Methods

  • Early theories of neural computation
  • Development of electrophysiological recording techniques
  • Impact of computer science on neural modeling

Section 2. Key Figures and Landmark Studies

  • Hodgkin and Huxley: Action potential modeling
  • Hebb: Synaptic plasticity and learning theory
  • Marr: Theories of cerebellar and hippocampal function

Chapter 3. Biological Foundations of Neural Systems

Section 1. Overview of Neuroanatomy

  • The structure and function of neurons
  • Basic components of the nervous system
  • Types of neurons and their specialized functions

Section 2. Cellular and Molecular Neuroscience Essentials

  • Ion channels, neurotransmitters, and synaptic transmission
  • Genetic and epigenetic influences on neural activity
  • Neurodevelopment and neural plasticity

Chapter 4. Neuroanatomy and Functional Organization of the Brain

Section 1. Structural Organization of Neural Circuits

  • Functional roles of different brain regions
  • Cortical layers and their connectivity patterns
  • Subcortical structures and their computational functions

Section 2. Regional Specialization and Brain Connectivity

  • Hierarchical processing in sensory and motor pathways
  • Large-scale network interactions and functional hubs
  • Role of white matter tracts and connectomics

Chapter 5. Mathematical Tools for Neural Modeling

Section 1. Foundations of Differential Equations in Neuroscience

  • Ordinary and partial differential equations in neural dynamics
  • Applications in modeling membrane potentials
  • Stability analysis and equilibrium points

Section 2. Linear Algebra and Matrix Methods

  • Eigenvalues, eigenvectors, and their neural interpretations
  • Principal component analysis (PCA) in neural population activity
  • Matrix operations in recurrent neural networks

Section 3. Calculus in Neural Computation

  • Gradient-based learning and optimization in neural models
  • Differential operators in continuous-time models
  • Stochastic calculus in neuronal signal processing

Section 4. Probability, Statistics, and Stochastic Processes

  • Gaussian processes and probability distributions in neural systems
  • Markov chains and hidden Markov models in neural state transitions
  • Monte Carlo methods in Bayesian inference for neuroscience

Chapter 6. Dynamical Systems in Neuroscience

Section 1. Introduction to Nonlinear Dynamics in Neural Systems

  • Fixed points, limit cycles, and attractors in neural activity
  • Chaos and complexity in brain dynamics
  • Stability criteria for neuronal systems

Section 2. Phase Space Representation and Bifurcation Theory

  • Phase portraits of neural oscillators
  • Bifurcations in neural excitability and bursting dynamics
  • Role of saddle-node and Hopf bifurcations in spiking activity

Section 3. Synchronization and Coupled Neural Oscillators

  • Mathematical models of coupled oscillators in the brain
  • Role of synchronization in cognition and perception
  • Kuramoto model and phase-locking phenomena in neural circuits

Chapter 7. Information Theory and Statistical Methods in Neural Computation

Section 1. Foundations of Information Theory in Neuroscience

  • Concepts of entropy, mutual information, and redundancy
  • Shannon information in neural spike trains
  • Neural coding efficiency and error correction mechanisms

Section 2. Bayesian Inference and Probabilistic Neural Models

  • Bayesian frameworks for sensory perception and learning
  • Hierarchical probabilistic models in decision-making
  • Markov decision processes and their applications in neuroscience

Section 3. Machine Learning and Statistical Learning in Neuroscience

  • Supervised, unsupervised, and reinforcement learning models
  • Applications of deep learning in neural decoding
  • Statistical methods for analyzing large-scale neural data

Chapter 8. Membrane Biophysics and Ion Channels

Section 1. Electrical Properties of Neurons

  • Resting membrane potential and ionic equilibrium
  • Role of capacitance and resistance in neuronal membranes
  • Passive vs. active membrane properties

Section 2. Ion Channel Kinetics and Membrane Dynamics

  • Voltage-gated ion channels: sodium, potassium, calcium
  • Channel conductance, gating variables, and activation/inactivation cycles
  • Stochastic ion channel models and their biological relevance

Section 3. Cable Theory and Passive Signal Transmission

  • Dendritic filtering and signal attenuation
  • Compartmental models for neuron morphology
  • Length constants and time constants in neural conduction

Chapter 9. The Hodgkin-Huxley Model and Beyond

Section 1. Detailed Examination of the Hodgkin-Huxley Equations

  • Development and biological basis of the model
  • Mathematical formulation of ionic currents
  • Simulating action potentials using Hodgkin-Huxley dynamics

Section 2. Simplified Models and Their Applications

  • FitzHugh-Nagumo model for reduced excitability representation
  • Integrate-and-fire models: leaky, quadratic, exponential variations
  • Izhikevich neuron model: balancing biological accuracy and efficiency

Section 3. Extensions of the Hodgkin-Huxley Framework

  • Multi-compartment models for complex neuron morphologies
  • Conductance-based models incorporating multiple ionic currents
  • Stochastic Hodgkin-Huxley models for single-neuron variability

Chapter 10. Neural Spiking, Action Potentials, and Signal Transmission

Section 1. Mechanisms of Action Potential Generation

  • Threshold dynamics and all-or-none principle
  • Refractory periods and spike adaptation
  • Influence of axonal geometry on conduction velocity

Section 2. Temporal Coding and Spike Train Analysis

  • Rate coding vs. temporal coding: implications for neural computation
  • Inter-spike intervals, burst coding, and Poisson statistics
  • Information-theoretic analysis of spike trains

Section 3. Propagation of Action Potentials in Neural Networks

  • Myelination and saltatory conduction in axons
  • Role of axonal delays in network synchronization
  • Computational models of spike transmission across brain regions

Chapter 11. Synaptic Transmission and Plasticity

Section 1. Chemical and Electrical Synapses

  • Mechanisms of neurotransmitter release and receptor activation
  • Gap junctions and direct electrical coupling in neural circuits
  • Differences in speed and reliability of synaptic transmission

Section 2. Mechanisms of Short-Term and Long-Term Plasticity

  • Short-term synaptic dynamics: facilitation, depression, and adaptation
  • Hebbian plasticity: synaptic strengthening and weakening rules
  • Role of spike-timing-dependent plasticity (STDP) in learning

Section 3. Molecular and Computational Mechanisms of Learning

  • Long-term potentiation (LTP) and long-term depression (LTD) mechanisms
  • Biochemical signaling pathways underlying synaptic modifications
  • Computational models of synaptic weight adaptation

Chapter 12. Single Neuron Models

Section 1. Integrate-and-Fire Models

  • Leaky integrate-and-fire model and its dynamics
  • Quadratic and exponential integrate-and-fire models
  • Limitations and extensions of integrate-and-fire models

Section 2. FitzHugh-Nagumo and Other Reduced Models

  • Simplifications of the Hodgkin-Huxley framework
  • Phase plane analysis and excitability classification
  • Biophysical relevance and computational efficiency

Section 3. Nonlinear and Hybrid Neuron Models

  • Izhikevich model and its ability to reproduce diverse spiking patterns
  • Adaptive exponential integrate-and-fire model
  • Role of fractional-order models in neural computation

Chapter 13. Feedforward and Recurrent Neural Networks

Section 1. Architecture and Dynamics of Feedforward Networks

  • Perceptron models and linear separability
  • Multi-layer networks and universal approximation theory
  • Role of synaptic weights and learning rules

Section 2. Role of Recurrent Connections in Memory and Pattern Generation

  • Auto-associative and hetero-associative memory models
  • Attractor networks and stability analysis
  • Computational role of reentry and feedback loops

Section 3. Cortical Circuit Models and Large-Scale Computation

  • Cortical microcircuit modeling approaches
  • Hierarchical processing and feature extraction in biological networks
  • Functional specialization and global integration in neural networks

Chapter 14. Oscillations, Synchronization, and Network Dynamics

Section 1. Rhythms in Neural Circuits

  • Alpha, beta, gamma, and theta oscillations in cognition
  • Origin and functional roles of brain rhythms
  • Relationship between oscillations and cognitive performance

Section 2. Mechanisms of Synchronization and Their Computational Roles

  • Phase-locking and coherence in neural activity
  • Synchronization in sensory processing and attention
  • Role of inhibitory interneurons in network synchronization

Section 3. Computational Models of Neural Synchrony

  • Kuramoto model and phase-coupled oscillators
  • Wilson-Cowan equations and population-level oscillations
  • Synchronization in spiking neural networks

Chapter 15. Artificial Neural Networks: Bridging Biological and Machine Learning Models

Section 1. Comparison of Biological Neural Networks with Artificial Networks

  • Differences in learning paradigms: supervised vs. unsupervised learning
  • Neuromodulation and synaptic plasticity in artificial models
  • Computational efficiency vs. biological realism trade-offs

Section 2. Deep Learning and Its Inspiration from Neuroscience

  • Hierarchical feature learning in deep networks vs. biological vision
  • Recurrent and convolutional neural networks in cognitive tasks
  • Neuromorphic engineering and biologically plausible architectures

Section 3. Spiking Neural Networks and Energy-Efficient Computation

  • Temporal coding in spiking neural networks
  • Hardware implementations of biologically inspired networks
  • Role of neuromorphic computing in next-generation AI

Chapter 16. Sensory Processing and Neural Coding

Section 1. Models of Visual, Auditory, and Somatosensory Systems

  • Retinotopic, tonotopic, and somatotopic organization in the brain
  • Hierarchical processing in the visual cortex
  • Computational models of sound localization and tactile perception

Section 2. Encoding and Decoding Strategies

  • Rate coding, temporal coding, and sparse coding theories
  • Bayesian inference in sensory perception
  • Reverse correlation and decoding population activity

Section 3. Predictive Coding and Perception

  • Hierarchical predictive processing in sensory systems
  • Role of feedback loops in perception and cognition
  • Computational models of expectation and surprise in perception

Chapter 17. Motor Control and Movement Generation

Section 1. Computational Models of Motor Systems

  • Role of the motor cortex, basal ganglia, and cerebellum in movement
  • Forward and inverse models of motor planning
  • Neural representations of movement trajectories

Section 2. Feedback Control and Sensorimotor Integration

  • Optimal control and reinforcement learning in movement execution
  • Proprioceptive feedback and its role in coordination
  • Computational models of sensorimotor adaptation

Section 3. Brain-Machine Interfaces for Motor Control

  • Decoding movement intentions from neural signals
  • Real-time control of prosthetics and robotic limbs
  • Future directions in brain-computer interface research

Chapter 18. Cognitive Processes: Learning, Memory, and Decision Making

Section 1. Theories and Models of Memory Formation

  • Hippocampal models of episodic and spatial memory
  • Computational models of working memory in the prefrontal cortex
  • Long-term memory consolidation and synaptic reorganization

Section 2. Computational Frameworks for Decision-Making Processes

  • Reinforcement learning and value-based decision-making
  • Bayesian decision models and probabilistic reasoning
  • The role of uncertainty and confidence estimation in cognition

Section 3. Executive Function and Cognitive Control

  • Neural mechanisms of attention switching and inhibitory control
  • Hierarchical models of goal-directed behavior
  • Computational models of planning and problem-solving

Chapter 19. Attention, Perception, and Consciousness

Section 1. Modeling Attentional Mechanisms

  • Top-down vs. bottom-up attention models
  • Neural correlates of selective attention and feature binding
  • Computational models of visual search and attentional shifts

Section 2. The Computational Basis of Perception

  • Multisensory integration and cross-modal processing
  • Neural mechanisms of perceptual decision-making
  • Role of Bayesian inference in perceptual stability

Section 3. Theories of Consciousness in Computational Neuroscience

  • Global workspace theory and integrated information theory
  • Neural correlates of consciousness in the brain
  • Computational approaches to self-awareness and metacognition

Chapter 20. Data Acquisition in Neuroscience: Techniques and Technologies

Section 1. Neuroimaging, Electrophysiology, and Optical Methods

  • Functional MRI (fMRI), PET, and EEG for brain activity mapping
  • Invasive techniques: single-unit recording and ECoG
  • Optical imaging methods (two-photon microscopy, calcium imaging)

Section 2. Challenges in Data Collection and Preprocessing

  • Signal-to-noise ratio and artifacts in neurophysiological recordings
  • Motion correction and preprocessing pipelines for neuroimaging data
  • Ethical and practical constraints in human and animal studies

Section 3. Advances in Large-Scale Neural Data Acquisition

  • High-density electrode arrays and Neuropixels probes
  • Whole-brain recording techniques in model organisms
  • Emerging tools for high-throughput neurophysiology

Chapter 21. Neural Signal Processing and Analysis

Section 1. Time Series Analysis and Frequency Domain Methods

  • Power spectral density analysis and event-related potentials
  • Wavelet transforms for neural oscillation analysis
  • Signal filtering, detrending, and preprocessing techniques

Section 2. Machine Learning Applications in Neural Data Interpretation

  • Dimensionality reduction (PCA, ICA, t-SNE) in neural datasets
  • Clustering and classification of neural spiking patterns
  • Deep learning approaches for brain activity decoding

Section 3. Statistical Methods for Neural Data Analysis

  • Bayesian methods and probabilistic modeling in neuroscience
  • Multivariate pattern analysis (MVPA) for neuroimaging data
  • Hidden Markov models for neural state transitions

Chapter 22. Simulation Techniques and Software Tools

Section 1. Overview of Simulation Environments (e.g., NEURON, NEST, Brian)

  • Features and applications of major neural simulation tools
  • Model construction and parameter tuning
  • Case studies of large-scale brain simulations

Section 2. High-Performance Computing and Modeling Frameworks

  • Parallel computing and GPU acceleration in neural simulations
  • Cloud-based and distributed computing for large-scale models
  • Trade-offs between biophysical accuracy and computational efficiency

Section 3. Reproducibility and Standards in Neural Simulations

  • Model validation and parameter sensitivity analysis
  • Open-source datasets and collaborative research frameworks
  • Standardized formats for neural simulations (NeuroML, SONATA)

Chapter 23. Modeling Brain Connectivity and Network Analysis

Section 1. Graph Theory and Network Science in Neuroscience

  • Small-world and scale-free properties of brain networks
  • Community structure and modularity in neural networks
  • Centrality measures and their relevance to brain function

Section 2. Connectomics and the Analysis of Large-Scale Neural Networks

  • Structural vs. functional connectivity in brain mapping
  • Diffusion tensor imaging (DTI) and tractography
  • Machine learning approaches for network-based brain decoding

Section 3. Computational Models of Large-Scale Brain Activity

  • Whole-brain modeling approaches (e.g., Virtual Brain)
  • Dynamic functional connectivity and brain state transitions
  • Predictive modeling of cognitive functions using network neuroscience

Chapter 24. Neuromorphic Engineering and Brain-Inspired Hardware

Section 1. Principles of Neuromorphic Design

  • Analog vs. digital neuromorphic circuits
  • Event-driven computation and energy efficiency
  • Role of spike-timing-dependent plasticity (STDP) in hardware

Section 2. Applications and Case Studies

  • IBM TrueNorth and Intel Loihi architectures
  • Neuromorphic sensors for edge AI and robotics
  • Use cases in brain-machine interfaces and adaptive control

Section 3. Limitations and Future Prospects

  • Challenges in scaling neuromorphic systems
  • Comparison with traditional deep learning accelerators
  • Potential for hybrid AI-neuroscience architectures

Chapter 25. The Intersection of AI and Computational Neuroscience

Section 1. How Neuroscience Informs Artificial Intelligence

  • Biological learning principles adapted to AI models
  • Computational role of memory and attention in deep learning
  • Hebbian learning vs. backpropagation in artificial networks

Section 2. Cross-Fertilization Between Deep Learning and Neural Modeling

  • Predictive coding and hierarchical Bayesian models
  • Reinforcement learning insights from dopamine-based circuits
  • Spiking neural networks (SNNs) as an alternative to ANN architectures

Section 3. Future Directions in Brain-Inspired AI

  • Bridging symbolic AI and neural computation
  • Neuromorphic computing for real-time AI applications
  • Ethical considerations in AI systems modeled after the brain

Chapter 26. Computational Approaches to Neurological Disorders

Section 1. Modeling Disease Mechanisms and Neural Dysfunction

  • Computational models of epilepsy and seizure dynamics
  • Schizophrenia as a disorder of disrupted neural synchrony
  • Neurodegenerative diseases and computational aging models

Section 2. Applications in Diagnosis and Therapeutic Interventions

  • Machine learning for early detection of neurological disorders
  • Computational biomarkers in neuropsychiatric conditions
  • Brain stimulation techniques: TMS, DBS, and neurofeedback

Section 3. Predictive Medicine and Personalized Neuroscience

  • AI-driven personalized treatment strategies
  • Simulating drug effects using computational brain models
  • Ethical considerations in predictive neuroscience

Chapter 27. Ethical Considerations and the Future of Brain Research

Section 1. Ethical Issues in Computational Modeling and Neurotechnology

  • Data privacy concerns in large-scale neural recording
  • Potential misuse of neurotechnology in surveillance and warfare
  • Bias and fairness in AI-based brain research

Section 2. Prospects and Challenges for the Future

  • The role of open science in computational neuroscience
  • Implications of whole-brain simulation projects
  • The future of human-AI brain augmentation

Section 3. Philosophical Questions and Theoretical Frontiers

  • Can computational models fully explain consciousness?
  • The limits of reductionism in neuroscience
  • Ethical dilemmas in brain emulation and mind uploading

Chapter 28. Mathematical Appendices and Tutorials

Section 1. Detailed Derivations and Additional Mathematical Background

  • Step-by-step derivations of key computational models
  • Linear algebra review: eigenvalues, eigenvectors, and matrices
  • Differential equations and stability analysis in neural systems

Section 2. Worked Examples and Exercises

  • Solving the Hodgkin-Huxley equations numerically
  • Implementing neural network models from first principles
  • Exercises in Bayesian inference for neural coding

Section 3. Advanced Mathematical Tools for Neuroscience

  • Fourier analysis for neural signal processing
  • Optimization methods in neural network training
  • Information-theoretic approaches to spike train analysis

Chapter 29. Programming Examples and Simulation Code

Section 1. Sample Code Snippets in Python, MATLAB, or Other Relevant Languages

  • Implementing integrate-and-fire neuron models
  • Simulating large-scale neural networks in NEURON and NEST
  • Computational models of synaptic plasticity using Python

Section 2. Guides to Setting Up and Running Simulations

  • Installing and configuring neural simulation software
  • Running large-scale network models on high-performance computing clusters
  • Using Jupyter notebooks for interactive computational neuroscience

Section 3. Best Practices in Computational Modeling

  • Reproducibility and documentation of neural simulations
  • Debugging and optimizing large-scale brain models
  • Version control and collaborative workflows for neuroscience research

Chapter 30. Glossary, Further Reading, and Online Resources

Section 1. Definitions of Key Terms and Concepts

  • Comprehensive glossary of computational neuroscience terminology
  • Explanation of essential mathematical and biological terms
  • Quick-reference guide to common modeling frameworks

Section 2. Annotated Bibliography and Recommendations for Further Exploration

  • Foundational textbooks and landmark papers in computational neuroscience
  • Online courses, tutorials, and video lectures
  • Recommended software tools and repositories for neural modeling

Section 3. Additional Resources and Community Involvement

  • Open-source datasets and neural simulation platforms
  • Conferences and journals dedicated to computational neuroscience
  • How to contribute to neuroscience research and open science initiatives

The Leanpub 60 Day 100% Happiness Guarantee

Within 60 days of purchase you can get a 100% refund on any Leanpub purchase, in two clicks.

Now, this is technically risky for us, since you'll have the book or course files either way. But we're so confident in our products and services, and in our authors and readers, that we're happy to offer a full money back guarantee for everything we sell.

You can only find out how good something is by trying it, and because of our 100% money back guarantee there's literally no risk to do so!

So, there's no reason not to click the Add to Cart button, is there?

See full terms...

Earn $8 on a $10 Purchase, and $16 on a $20 Purchase

We pay 80% royalties on purchases of $7.99 or more, and 80% royalties minus a 50 cent flat fee on purchases between $0.99 and $7.98. You earn $8 on a $10 sale, and $16 on a $20 sale. So, if we sell 5000 non-refunded copies of your book for $20, you'll earn $80,000.

(Yes, some authors have already earned much more than that on Leanpub.)

In fact, authors have earned over $14 million writing, publishing and selling on Leanpub.

Learn more about writing on Leanpub

Free Updates. DRM Free.

If you buy a Leanpub book, you get free updates for as long as the author updates the book! Many authors use Leanpub to publish their books in-progress, while they are writing them. All readers get free updates, regardless of when they bought the book or how much they paid (including free).

Most Leanpub books are available in PDF (for computers) and EPUB (for phones, tablets and Kindle). The formats that a book includes are shown at the top right corner of this page.

Finally, Leanpub books don't have any DRM copy-protection nonsense, so you can easily read them on any supported device.

Learn more about Leanpub's ebook formats and where to read them

Write and Publish on Leanpub

You can use Leanpub to easily write, publish and sell in-progress and completed ebooks and online courses!

Leanpub is a powerful platform for serious authors, combining a simple, elegant writing and publishing workflow with a store focused on selling in-progress ebooks.

Leanpub is a magical typewriter for authors: just write in plain text, and to publish your ebook, just click a button. (Or, if you are producing your ebook your own way, you can even upload your own PDF and/or EPUB files and then publish with one click!) It really is that easy.

Learn more about writing on Leanpub