Saturday, March 22, 2025

AutoDock Vina: A Comprehensive Overview

AutoDock Vina: A Comprehensive Overview

1. Introduction

AutoDock Vina is a widely used molecular docking software designed for predicting the binding affinity and binding poses of small molecules (ligands) with target proteins (receptors). It is an improved version of the original AutoDock software and is known for its enhanced accuracy and significantly faster performance.

AutoDock Vina is particularly popular in the fields of drug discovery, computational chemistry, and structural biology. It is open-source and developed by The Scripps Research Institute.


2. Key Features

  • High Speed and Accuracy:

    • AutoDock Vina is much faster than its predecessor, AutoDock 4, due to its efficient scoring function and optimization algorithms.
    • Provides more reliable docking results with better pose prediction.
  • Simple and Automated Workflow:

    • Requires minimal user intervention and is easier to set up than AutoDock.
    • Automates many parameter settings, making it user-friendly.
  • Flexible Ligand and Receptor Docking:

    • Supports rigid docking (fixed receptor, flexible ligand) and flexible receptor docking (selected residues flexible).
  • Multi-Core CPU Support:

    • Can utilize multiple processor cores to speed up calculations.
  • Energy-Based Scoring Function:

    • Uses an empirical scoring function to estimate binding affinity in kcal/mol.
  • Wide Compatibility:

    • Compatible with Linux, Windows, and macOS.
    • Works well with AutoDockTools (ADT) for input file preparation and visualization.

3. How AutoDock Vina Works

AutoDock Vina performs molecular docking by following these steps:

  1. Protein and Ligand Preparation:

    • The receptor (protein) and ligand structures are prepared in PDBQT format using AutoDockTools (ADT).
    • The receptor is usually kept rigid, while the ligand is assigned rotatable bonds.
  2. Defining the Search Space (Grid Box):

    • A search box is defined around the active site of the receptor, specifying where the ligand can explore binding conformations.
  3. Docking Process:

    • The software uses an iterative global-local optimization algorithm to generate ligand conformations and predict binding poses.
    • The binding energy of each pose is calculated using its scoring function.
  4. Result Analysis:

    • The best docking pose (lowest binding energy) is selected.
    • Users analyze the output PDBQT files using visualization tools like PyMOL, Chimera, or Discovery Studio.

4. Applications

  • Drug Discovery:

    • Identifying lead compounds by screening molecular libraries.
    • Predicting drug-receptor interactions.
  • Enzyme Inhibitor Design:

    • Modeling how small molecules inhibit enzymes by binding to active sites.
  • Protein-Ligand Interaction Studies:

    • Understanding molecular interactions to aid in rational drug design.
  • Virtual Screening:

    • Screening large compound libraries to find potential drug candidates.

5. Advantages & Limitations

Advantages:
  • Free and open-source.
  • Faster than AutoDock 4.
  • User-friendly and requires minimal setup.
  • Supports parallel computing (multi-threading).
  • Provides accurate binding energy predictions.
Limitations:
  • Cannot handle covalent docking directly.
  • Less flexible receptor handling than more advanced tools like RosettaDock.
  • Limited to rigid body docking with only selected receptor flexibility.

6. Comparison: AutoDock Vina vs. AutoDock 4

Feature AutoDock Vina AutoDock 4
Speed Faster Slower
Scoring Function Empirical Grid-based
Ease of Use Easier More complex
Multi-threading Yes No
Flexible Receptor Limited More control

7. Getting Started with AutoDock Vina

Installation
  • Download from the official AutoDock website.
  • Available for Windows, Linux, and macOS.
  • Requires Python, OpenBabel, and AutoDockTools for file preparation.
Basic Command Line Usage
vina --receptor protein.pdbqt --ligand ligand.pdbqt --center_x 10 --center_y 20 --center_z 15 --size_x 20 --size_y 20 --size_z 20 --out output.pdbqt

This command specifies:

  • The receptor and ligand files.
  • The docking grid center and size.
  • The output file containing predicted poses.

8. Related Tools for Visualization

  • PyMOL – View and analyze docked complexes.
  • Chimera – Advanced molecular visualization.
  • Discovery Studio – Commercial tool with detailed interaction analysis.

Conclusion

AutoDock Vina is a powerful
, free, and efficient docking tool widely used in computational drug discovery. Its ease of use, speed, and improved scoring function make it a preferred choice over AutoDock 4 for many researchers.

to download it:
https://vina.scripps.edu/

To get a consultancy on your new docking project, please contact me at MPA@pharmakoi.com


Enjoy!!

Mass




Sunday, March 16, 2025

the ASUS TUF Gaming B850-PLUS WIFI motherboard

The Observer Corner:

Today we dive into the ASUS TUF Gaming B850-PLUS WIFI motherboard, one of the best price/performance ration motherboard in my personal opinion.

The ASUS TUF Gaming B850-PLUS WIFI motherboard is an ATX board designed for AMD Ryzen 9000, 8000, and 7000 series processors. It features PCIe 5.0 x16 support, Wi-Fi 7, and Realtek 2.5Gb Ethernet, making it ideal for gaming and high-performance computing.

Key Specifications:

  • Graphics Outputs: DisplayPort (8K@30Hz) and HDMI 2.1 (4K@60Hz).
  • Expansion Slots: PCIe 5.0 (x16), PCIe 4.0 (x16, x8/x4 mode), and PCIe 4.0 (x4, x1 slots).
  • Storage: 3x M.2 slots (PCIe 5.0/4.0) and 4x SATA 6Gb/s ports.
  • USB Ports:
    • Rear I/O: 1x USB-C (20Gbps), 3x USB-A (10Gbps), 4x USB-A (5Gbps), and 2x USB 2.0.
    • Front Panel: 1x USB-C (10Gbps), 2x USB 5Gbps, and 4x USB 2.0.
  • Networking: Wi-Fi 7 (up to 2.9Gbps) and Bluetooth 5.4.
  • Audio: Realtek ALC1220P 7.1 Surround Sound with premium audio components.
  • Cooling & Power: 4+ chassis fan headers, 1x AIO pump header, 2x 8-pin CPU power connectors.

This motherboard includes ASUS TUF PROTECTION, Q-Design features for easy installation, and Aura Sync RGB headers for customization.

Take a look at the link below for more details:
https://dlcdnets.asus.com/pub/ASUS/mb/SocketAM5/TUF_GAMING_B850-PLUS_WIFI/E25809_TUF_GAMING_B850-PLUS_WIFI_UM_V2_WEB.pdf?model=TUF%20GAMING%20B850-PLUS%20WIFI

Enjoy!!

Massimiliano
Perugia, March 15th, 2025



Latest trends in GPU technology

Perugia - March 9th, 2025


The latest trends in GPU technology for fluid simulation highlight significant advancements in performance, scalability, and cost efficiency.

GPU Acceleration in Computational Fluid Dynamics (CFD)
GPUs are now an essential tool in CFD, drastically reducing simulation times. Tasks that once took an entire day on CPU servers can now be completed in just over an hour using multiple high-performance GPUs. This acceleration benefits industries such as aerospace, automotive, and pharmaceuticals, where fluid dynamics simulations play a critical role in research and development.

Scalability and Multi-GPU Configurations
Multi-GPU setups are becoming more prevalent, offering improved computational power and efficiency. FluidX3D, for example, has demonstrated a system combining Intel and NVIDIA GPUs to maximize performance while keeping costs lower than high-end single-GPU solutions. The ability to integrate GPUs from different vendors allows for more flexible and cost-effective simulation environments.

Optimized GPU Selection for Specific Workloads
Choosing the right GPU depends on the simulation requirements. Consumer-grade GPUs like the RTX 4090 are excellent for single-precision workloads, providing high performance at a lower cost. On the other hand, enterprise GPUs such as the NVIDIA H100 and A100 excel in handling double-precision and memory-intensive tasks, making them more suitable for large-scale and highly detailed simulations.

Cloud and Hybrid Deployments
Many CFD software providers, including industry leaders like Ansys and Siemens, are optimizing their tools for GPU acceleration in both on-premise and cloud-based environments. Cloud solutions powered by high-performance GPUs enable scalable, on-demand simulations, reducing infrastructure costs and increasing accessibility for researchers and engineers.

Expansion of Competition in High-Performance CFD
AMD is making strides in the high-performance computing space with its Instinct MI300X GPU, which is specifically designed to handle computationally heavy simulations. This competition provides more options for researchers and engineers, challenging NVIDIA’s dominance in the field and fostering further innovation.

Overall, GPUs are transforming fluid simulation by making it faster, more efficient, and more scalable. With continued advancements in hardware and software optimization, the future of CFD looks increasingly driven by high-performance GPU computing.


Interested to a custom-built workstation?
Send out your inquiry to MPA@pharmakoi.com indicating the overall performances you are looking for (TFlops, etc...) and you will get a free quote of a proposed configuration.





Saturday, January 4, 2025

why Nvidia A40 GPUs are so popular?

The NVIDIA A40 GPU is popular for its versatility and high performance across various computational workloads. Here’s why it stands out:


1. Designed for Versatile Use

The NVIDIA A40 is built to handle diverse workloads, including:

  • AI and Machine Learning: Its architecture supports AI training and inference with high precision.
  • Graphics Rendering: Offers exceptional rendering capabilities for virtual environments and 3D applications.
  • High-Performance Computing (HPC): Optimized for computational tasks like simulations, scientific research, and cryptocurrency mining.

This flexibility makes the A40 appealing across industries, from AI research to creative design and enterprise workloads.


2. Ampere Architecture

The A40 is based on NVIDIA's Ampere architecture, which includes:

  • CUDA Cores: A significant number of CUDA cores (10,752) to accelerate parallel processing tasks.
  • RT Cores and Tensor Cores: Enhancements for ray tracing and AI-specific operations.
  • Memory Bandwidth: Equipped with 48GB of GDDR6 memory and a bandwidth of 696 GB/s, making it ideal for memory-intensive applications.

These architectural advancements provide a significant performance boost over previous generations, contributing to its popularity.


3. Excellent Performance-to-Cost Ratio

Compared to flagship GPUs like the NVIDIA A100, the A40 provides excellent computational and rendering performance at a relatively lower price point. This balance between performance and cost makes it attractive for enterprises looking for powerful solutions without overspending.


4. Enterprise and Data Center Optimizations

  • Passive Cooling Design: Designed for data center environments, the A40 has a passive cooling mechanism, making it ideal for server racks.
  • Virtualization: Supports NVIDIA’s virtual GPU (vGPU) technology, enabling use cases in virtual desktops and high-performance rendering in remote environments.

5. Popular in Cryptocurrency Mining

The A40 has gained popularity among cryptocurrency miners due to its:

  • High Hash Rates: Especially for memory-intensive algorithms like Ethereum before the shift to proof-of-stake.
  • Energy Efficiency: Provides a good balance of performance per watt, which is critical for mining profitability.

6. Preferred for AI and HPC

  • AI Training: Its Tensor Cores enable efficient processing of AI workloads, while its large memory capacity supports large models and datasets.
  • Inference: With mixed-precision capabilities, it can handle real-time AI inference tasks effectively.
  • HPC Applications: Its ability to process complex scientific computations makes it a favored choice in research and enterprise HPC environments.

7. Industry Adoption and Ecosystem

  • Widely supported in major deep learning and HPC frameworks like TensorFlow, PyTorch, and MATLAB.
  • Integrated into cloud services and enterprise solutions, making it accessible to a broader range of users.

The NVIDIA A40 GPU’s combination of advanced architecture, diverse use cases, and a competitive performance-to-cost ratio makes it a popular choice across sectors like AI, HPC, graphics rendering, and cryptocurrency mining.




Thursday, January 2, 2025

a detailed technical comparison of Ubuntu and CentOS, focusing on aspects relevant to computational tasks and industrial use cases

1. Base and Philosophy

  • Ubuntu:
    • Base: Debian-based.
    • Philosophy: Prioritizes usability, regular updates, and a large ecosystem. Ideal for both desktop and server environments.
    • Target Users: Developers, researchers, and users looking for a balance of cutting-edge and stability.
  • CentOS:
    • Base: Historically based on Red Hat Enterprise Linux (RHEL). After CentOS Stream's introduction, it now serves as RHEL's upstream.
    • Philosophy: Stability and predictability. Ideal for enterprise environments needing long-term support and tested packages.
    • Target Users: Enterprises requiring rock-solid stability and HPC clusters.

2. Package Management

  • Ubuntu:
    • Package Manager: APT (Advanced Package Tool), which uses .deb packages.
    • Repositories: Includes Main, Universe, Restricted, and Multiverse repositories, offering a large selection of pre-built software.
    • Advantages:
      • Faster updates and access to newer software versions.
      • Strong focus on compatibility with modern software (e.g., Python, machine learning libraries).
  • CentOS:
    • Package Manager: YUM or DNF (on newer versions), which uses .rpm packages.
    • Repositories: Limited compared to Ubuntu by default, but extended using EPEL (Extra Packages for Enterprise Linux) and third-party repos.
    • Advantages:
      • Highly stable, enterprise-ready software versions.
      • Better suited for systems requiring strict version control (e.g., older Python or GCC for compatibility).

3. Release Cycle and Updates

  • Ubuntu:

    • Releases: Two versions:
      • LTS (Long-Term Support): Released every two years, supported for 5 years (e.g., 20.04, 22.04).
      • Non-LTS: Released every six months, supported for 9 months.
    • Update Frequency: Frequent updates with newer features, kernels, and software versions.
    • Best Use: Projects needing cutting-edge software and hardware support.
  • CentOS:

    • Releases:
      • CentOS Stream: Continuous updates as the upstream development version of RHEL.
      • CentOS 7/8 Legacy: Provided stability-focused updates, now largely replaced by CentOS Stream, AlmaLinux, or Rocky Linux.
    • Update Frequency: Slower and more deliberate updates focused on stability.
    • Best Use: Environments requiring long-term stability with minimal changes.

4. System Performance

  • Ubuntu:
    • Kernel: Ships with relatively new kernels in both LTS and non-LTS versions, allowing better hardware compatibility.
    • Performance: Optimized for modern workloads but may introduce slight instability due to newer software versions.
    • System Overhead: Lightweight flavors like Ubuntu Server or Ubuntu Minimal reduce overhead.
  • CentOS:
    • Kernel: Uses older, more stable kernel versions optimized for enterprise use. Hardware enablement may require backporting.
    • Performance: Focuses on consistency and low overhead in enterprise settings.
    • System Overhead: Minimal by design; better for high-load and mission-critical tasks.

5. Community and Enterprise Support

  • Ubuntu:

    • Community Support: Large and active community with extensive online documentation.
    • Enterprise Support: Canonical offers enterprise support for Ubuntu (e.g., Ubuntu Advantage).
    • Ecosystem: Widely used in machine learning, AI, and cloud environments like AWS and Azure.
  • CentOS:

    • Community Support: Smaller community compared to Ubuntu but still active in enterprise and HPC environments.
    • Enterprise Support: None directly for CentOS; instead, enterprises turn to RHEL, AlmaLinux, or Rocky Linux for support.
    • Ecosystem: Favored in HPC, scientific computing, and traditional enterprise environments.

6. Software Availability

  • Ubuntu:
    • Default Software: Supports a broader range of newer packages.
    • Compatibility: Better suited for modern languages, libraries, and frameworks (e.g., TensorFlow, Docker).
    • Cloud Integration: Leading choice for cloud-native technologies like Kubernetes and containerized applications.
  • CentOS:
    • Default Software: Ships with older, highly stable versions.
    • Compatibility: Ideal for legacy applications or systems requiring specific older software versions.
    • Cloud Integration: Supported but less prominent compared to Ubuntu.

7. HPC and Computational Workloads

  • Ubuntu:

    • Preferred for machine learning, AI, and development environments due to cutting-edge tools and frameworks.
    • Easier installation of GPU drivers (e.g., NVIDIA) and frameworks like TensorFlow or PyTorch.
  • CentOS:

    • Strong presence in HPC clusters and scientific computing.
    • Compatible with software requiring specific older libraries or system configurations.

8. Security and Compliance

  • Ubuntu:

    • Regular security updates.
    • Canonical provides enterprise-grade security solutions, including FIPS compliance.
    • Snap packages can introduce security concerns due to permissions model.
  • CentOS:

    • Stability-focused updates reduce the risk of security issues from newer software.
    • SELinux (Security-Enhanced Linux) is enabled by default, offering robust system security.

When to Use Ubuntu vs. CentOS

Feature Ubuntu CentOS
Modern Workloads Best for machine learning, AI, and cloud. Ideal for legacy or enterprise workloads.
Stability Moderate (LTS preferred). High (CentOS Stream or AlmaLinux).
Cutting-Edge Software Excellent. Limited; slower updates.
Long-Term Support 5 years (LTS). Enterprise-grade with RHEL.
Ease of Use Easier for beginners. Better for experienced admins.



Monday, December 30, 2024

OpenBabel online portal

OpenBabel online portal allows to convert nearly all the chemical formats and it very practical to quickly move from one program to another:
https://www.cheminfo.org/Chemistry/Cheminformatics/FormatConverter/index.html

Open Babel is a chemical toolbox designed to speak the many languages of chemical data. It’s an open, collaborative project allowing anyone to search, convert, analyze, or store data from molecular modeling, chemistry, solid-state materials, biochemistry, or related areas.

The latest version of this documentation is available in several formats from:

https://openbabel.org/docs/index.html

Enjoy!!








Saturday, December 28, 2024

How to estimate ADMET properties of a new drug

Estimating ADMET properties (Absorption, Distribution, Metabolism, Excretion, and Toxicity) is critical in drug discovery and development to predict the behavior of a drug in vivo. Here’s a detailed approach to estimating these properties:


1. Absorption

Absorption evaluates how effectively a drug enters systemic circulation after administration.

Key Factors:

  • Solubility: Affects dissolution rate and bioavailability.
  • Permeability: Indicates how well a drug crosses biological membranes.
  • pKa: Determines ionization state, which influences absorption in different pH environments.

Methods of Estimation:

  1. Experimental Techniques:

    • Caco-2 Cell Assay: Simulates intestinal epithelial permeability.
    • Parallel Artificial Membrane Permeability Assay (PAMPA): Assesses passive permeability.
    • Solubility Tests: Conducted in various pH buffers mimicking gastrointestinal conditions.
  2. In Silico Models:

    • Lipinski’s Rule of Five: Predicts oral bioavailability using molecular weight, hydrogen bond donors/acceptors, and lipophilicity (logP).
    • QSAR Models (Quantitative Structure-Activity Relationships): Relate molecular features to absorption data.
    • GastroPlus®: Simulates gastrointestinal absorption.

2. Distribution

Distribution assesses how a drug disperses throughout the body’s tissues and fluids.

Key Factors:

  • Volume of Distribution (Vd): Indicates the extent of drug distribution.
  • Plasma Protein Binding (PPB): Impacts free drug availability.
  • Tissue Binding: Influences drug accumulation in specific organs.

Methods of Estimation:

  1. Experimental Techniques:

    • Equilibrium Dialysis or Ultrafiltration: Measures plasma protein binding.
    • Animal Studies: Assess tissue-specific concentrations.
  2. In Silico Models:

    • LogP and LogD Calculations: Predict lipophilicity, a key determinant of tissue affinity.
    • Predictive Algorithms (e.g., pkCSM, ADMET Predictor): Estimate Vd and PPB from chemical structure.

3. Metabolism

Metabolism evaluates how a drug is chemically modified by enzymes, primarily in the liver.

Key Factors:

  • Phase I Metabolism: Involves oxidation, reduction, or hydrolysis (e.g., by cytochrome P450 enzymes).
  • Phase II Metabolism: Involves conjugation reactions like glucuronidation or sulfation.
  • Metabolic Stability: Reflects the drug’s half-life in metabolic systems.

Methods of Estimation:

  1. Experimental Techniques:

    • Liver Microsomes or Hepatocytes: Assess enzyme-mediated metabolism.
    • Cytochrome P450 Inhibition/Induction Studies: Identify potential drug-drug interactions.
    • Metabolite Identification: Using LC-MS/MS or NMR spectroscopy.
  2. In Silico Models:

    • SMARTCyp: Predicts likely metabolic sites.
    • MetaSite and StarDrop: Simulate enzyme-substrate interactions.
    • Physiologically-Based Pharmacokinetic (PBPK) Models: Estimate drug clearance and metabolic pathways.

4. Excretion

Excretion assesses how the drug and its metabolites are eliminated from the body.

Key Factors:

  • Renal Excretion: Includes glomerular filtration, tubular secretion, and reabsorption.
  • Biliary Excretion: Drug elimination via bile.
  • Half-life (t½): Indicates the duration of drug action.

Methods of Estimation:

  1. Experimental Techniques:

    • In Vitro Transporter Assays: Evaluate interaction with renal and hepatic transporters (e.g., OATs, OCTs, P-gp).
    • Animal Models: Measure urinary and fecal excretion.
  2. In Silico Models:

    • Clearance Predictions: Based on molecular size, charge, and hydrophobicity.
    • Transporter Interaction Models: Predict transporter-mediated excretion pathways.

5. Toxicity

Toxicity predicts adverse effects that a drug may induce.

Key Factors:

  • Acute Toxicity: Evaluates single-dose lethality.
  • Chronic Toxicity: Assesses long-term exposure effects.
  • Organ-Specific Toxicity: Focus on liver, kidneys, and heart.
  • Genotoxicity: Risk of DNA damage.
  • Off-Target Effects: Interactions with unintended biological targets.

Methods of Estimation:

  1. Experimental Techniques:

    • Cytotoxicity Assays: Using cell lines to assess viability (e.g., MTT, LDH assays).
    • hERG Assays: Test potential for cardiac arrhythmias.
    • In Vivo Studies: Evaluate systemic toxicity in animal models.
  2. In Silico Models:

    • DEREK Nexus and ADMET Predictor: Assess structural alerts for toxicity.
    • T.E.S.T. (Toxicity Estimation Software Tool): Predicts various toxicity endpoints.
    • QSAR for Specific Toxicity: Models for mutagenicity, carcinogenicity, or reproductive toxicity.

Integrated Approaches for ADMET Estimation

  1. High-Throughput Screening (HTS):

    • Combines automated assays for multiple ADMET parameters.
    • Prioritizes compounds with favorable profiles.
  2. In Silico Workflow:

    • Use cheminformatics platforms like Schrödinger’s QikProp or ADMETlab for rapid screening of ADMET properties.
  3. PBPK Modeling:

    • Combines ADMET data with physiological parameters for holistic predictions of drug behavior in humans.
  4. Machine Learning Models:

    • Utilize datasets of known drugs to predict ADMET properties based on chemical descriptors.

Conclusion

By integrating experimental data with in silico predictions, researchers can efficiently estimate ADMET properties and prioritize promising drug candidates for further development. Balancing cost, speed, and accuracy is key to successful ADMET profiling.




Extraction of Chemical Data from Literature Using Large Language Models: Opportunities and Expertise

  Abstract The rapid expansion of chemical and pharmaceutical literature presents both an opportunity and a challenge: while vast amounts of...