Skip to content

TactileNet is an AI-powered framework for generating accessible tactile graphics using Stable Diffusion, enabling scalable production of embossing-ready templates for visually impaired individuals.

License

Notifications You must be signed in to change notification settings

Adnan-Khan7/TactileNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

30 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

TactileNet: Bridging the Accessibility Gap with AI-Generated Tactile Graphics for Individuals with Vision Impairment

Official repository for project: TactileNet
Authors: Adnan Khan, Alireza Choubineh, Mai A. Shaaban, Abbas Akkasi, Majid Komeili

Paper Status: Accepted and to be presented at IEEE SMC 2025. More info: https://www.ieeesmc2025.org/

Project Page License: MIT


🔍 Abstract

Tactile graphics are essential for providing access to visual information for the 43 million people globally living with vision loss. Traditional methods for creating these graphics are labor-intensive and cannot meet growing demand. We introduce TactileNet, the first comprehensive dataset and AI-driven framework for generating embossing-ready 2D tactile templates using text-to-image Stable Diffusion models. We fine-tune Stable Diffusion models using Low-Rank Adaptation and DreamBooth to generate high-fidelity, guideline-compliant graphics with reduced computational cost. Quantitative evaluations with tactile experts show 92.86% adherence to accessibility standards. Our structural fidelity analysis revealed near-human design similarity, with a Structural Similarity Index (SSIM) of 0.538 between generated and expert-designed tactile images. Notably, our method better preserves object silhouettes than human designs (binary mask SSIM: 0.259 vs. 0.215), addressing a key limitation of manual abstraction. The framework scales to 32,000 images (7,050 high-quality) across 66 classes, with prompt editing enabling customizable outputs (e.g., adding or removing details). By automating the 2D template generation step—compatible with standard embossing workflows—TactileNet accelerates production while preserving design flexibility. This work demonstrates how AI can augment (not replace) human expertise to bridge the accessibility gap in education and beyond.

TL;DR: TactileNet uses small, domain-tuned LoRA + DreamBooth adapters to generate high-fidelity tactile graphics for visually impaired users using Stable Diffusion.


🎯 Teaser

Teaser Image

Figure: Our image-to-image translation pipeline. Top: natural reference images. Middle: benchmark tactile graphics from expert sources. Bottom: model-generated tactile graphics using our fine-tuned adapters.


🧠 Method Overview

TactileNet uses class-specific LoRA and DreamBooth adapters fine-tuned on 66 categories of tactile graphics. Each adapter is trained on a small number of high-quality, expert-sourced images and their structured prompts. The model supports both text-to-image and image-to-image generation via Stable Diffusion v1.5.

Key Features:

  • High SSIM similarity (0.538 vs expert designs)

  • Better silhouette preservation (0.259 vs 0.215 in binary SSIM)

  • Fully compatible with embossing workflows

  • Supports prompt-level editing (e.g., add/remove logo, features)

    TactileNet Methodology Diagram

Figure: Overview of the TactileNet framework combining LoRA and DreamBooth adapters for class-specific tactile image generation.


📂 Dataset

We release the full text-image TactileNet training dataset containing image-prompt pairs across 66 tactile graphic categories.

  • Organized by class folders (e.g., airplane/, bat/, etc.)
  • Each class folder includes an Inputs/ subfolder with:
    • Image files (.png or .jpg)
    • Corresponding prompt text files (.txt) containing natural language descriptions
  • Suitable for fine-tuning text-to-image or image-to-image generative models

📥 Access the dataset on Hugging Face:
👉 MaiAhmed/TactileNet (train split)


⚙️ Setup Instructions

1. Set Up Stable Diffusion WebUI

Follow instructions from the Stable Diffusion WebUI repository.

2. Clone Repository

git clone https://github.com/Adnan-Khan7/TactileNet.git
cd TactileNet

3. Copy Adapters

Download LoRA adapters from Google Drive

Copy all the .safetensors files to:

stable-diffusion-webui/models/Lora/

4. Download Base Models

  • SD v1.5: Get v1-5-pruned-emaonly.safetensors from Hugging Face.
  • Deliberate v3 (optional): Download

Place in: /stable-diffusion-webui/models/Stable-diffusion/


🧪 Usage & Configuration

  • Use the WebUI to generate tactile graphics with prompts like:

    "Create a tactile graphic of an airplane for visually impaired users, emphasizing raised wings, fuselage, and tail."

  • For image-to-image translation: Use denoising strength between 0.85–0.9
  • For text-to-image generation: Default settings (CFG=7, steps=20) suffice
  • Example configuration shown below:

Example


📈 Evaluation Protocol

  • 132 tactile graphics evaluated (66 generated + 66 benchmark)
  • 66 natural images as reference
  • Evaluation Metrics:
    • Pose accuracy (Q1)
    • Guideline adherence (Q2)
    • Expert-rated quality (Q3)
    • SSIM (structural fidelity)

📦 Reproducibility

Includes:

  • ssim.py for evaluating image similarity
  • Download the data for evaluation from Google Drive
  • Run ssim.py with data folder in base directory.

📄 License

This project is licensed under the MIT License.


🙏 Acknowledgments

This work was supported in part by MITACS, NSERC and the Digital Alliance of Canada. We thank the student volunteers at the Intelligent Machines Lab, Carleton University, for their help with dataset curation and image matching, especially Aarushi, Abrar, Lucksiha, Sohini, and Sona.

🚀 Citation

If you use TactileNet in your research, please cite:

@misc{khan2025tactilenetbridgingaccessibilitygap,
      title={TactileNet: Bridging the Accessibility Gap with AI-Generated Tactile Graphics for Individuals with Vision Impairment}, 
      author={Adnan Khan and Alireza Choubineh and Mai A. Shaaban and Abbas Akkasi and Majid Komeili},
      year={2025},
      eprint={2504.04722},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2504.04722}, 
}

About

TactileNet is an AI-powered framework for generating accessible tactile graphics using Stable Diffusion, enabling scalable production of embossing-ready templates for visually impaired individuals.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published