TactileNet: Bridging the Accessibility Gap with AI-Generated Tactile Graphics for Individuals with Vision Impairment
Official repository for project: TactileNet
Authors: Adnan Khan, Alireza Choubineh, Mai A. Shaaban, Abbas Akkasi, Majid Komeili
Paper Status: Accepted and to be presented at IEEE SMC 2025. More info: https://www.ieeesmc2025.org/
Tactile graphics are essential for providing access to visual information for the 43 million people globally living with vision loss. Traditional methods for creating these graphics are labor-intensive and cannot meet growing demand. We introduce TactileNet, the first comprehensive dataset and AI-driven framework for generating embossing-ready 2D tactile templates using text-to-image Stable Diffusion models. We fine-tune Stable Diffusion models using Low-Rank Adaptation and DreamBooth to generate high-fidelity, guideline-compliant graphics with reduced computational cost. Quantitative evaluations with tactile experts show 92.86% adherence to accessibility standards. Our structural fidelity analysis revealed near-human design similarity, with a Structural Similarity Index (SSIM) of 0.538 between generated and expert-designed tactile images. Notably, our method better preserves object silhouettes than human designs (binary mask SSIM: 0.259 vs. 0.215), addressing a key limitation of manual abstraction. The framework scales to 32,000 images (7,050 high-quality) across 66 classes, with prompt editing enabling customizable outputs (e.g., adding or removing details). By automating the 2D template generation step—compatible with standard embossing workflows—TactileNet accelerates production while preserving design flexibility. This work demonstrates how AI can augment (not replace) human expertise to bridge the accessibility gap in education and beyond.
TL;DR: TactileNet uses small, domain-tuned LoRA + DreamBooth adapters to generate high-fidelity tactile graphics for visually impaired users using Stable Diffusion.
Figure: Our image-to-image translation pipeline. Top: natural reference images. Middle: benchmark tactile graphics from expert sources. Bottom: model-generated tactile graphics using our fine-tuned adapters.
TactileNet uses class-specific LoRA and DreamBooth adapters fine-tuned on 66 categories of tactile graphics. Each adapter is trained on a small number of high-quality, expert-sourced images and their structured prompts. The model supports both text-to-image and image-to-image generation via Stable Diffusion v1.5.
Key Features:
-
High SSIM similarity (0.538 vs expert designs)
-
Better silhouette preservation (0.259 vs 0.215 in binary SSIM)
-
Fully compatible with embossing workflows
-
Supports prompt-level editing (e.g., add/remove logo, features)
Figure: Overview of the TactileNet framework combining LoRA and DreamBooth adapters for class-specific tactile image generation.
We release the full text-image TactileNet training dataset containing image-prompt pairs across 66 tactile graphic categories.
- Organized by class folders (e.g.,
airplane/
,bat/
, etc.) - Each class folder includes an
Inputs/
subfolder with:- Image files (
.png
or.jpg
) - Corresponding prompt text files (
.txt
) containing natural language descriptions
- Image files (
- Suitable for fine-tuning text-to-image or image-to-image generative models
📥 Access the dataset on Hugging Face:
👉 MaiAhmed/TactileNet (train split)
Follow instructions from the Stable Diffusion WebUI repository.
git clone https://github.com/Adnan-Khan7/TactileNet.git
cd TactileNet
Download LoRA adapters from Google Drive
Copy all the .safetensors files to:
stable-diffusion-webui/models/Lora/
- SD v1.5: Get
v1-5-pruned-emaonly.safetensors
from Hugging Face. - Deliberate v3 (optional): Download
Place in: /stable-diffusion-webui/models/Stable-diffusion/
- Use the WebUI to generate tactile graphics with prompts like:
"Create a tactile graphic of an airplane for visually impaired users, emphasizing raised wings, fuselage, and tail."
- For image-to-image translation: Use denoising strength between
0.85–0.9
- For text-to-image generation: Default settings (CFG=7, steps=20) suffice
- Example configuration shown below:
- 132 tactile graphics evaluated (66 generated + 66 benchmark)
- 66 natural images as reference
- Evaluation Metrics:
- Pose accuracy (Q1)
- Guideline adherence (Q2)
- Expert-rated quality (Q3)
- SSIM (structural fidelity)
Includes:
ssim.py
for evaluating image similarity- Download the data for evaluation from Google Drive
- Run ssim.py with data folder in base directory.
This project is licensed under the MIT License.
This work was supported in part by MITACS, NSERC and the Digital Alliance of Canada. We thank the student volunteers at the Intelligent Machines Lab, Carleton University, for their help with dataset curation and image matching, especially Aarushi, Abrar, Lucksiha, Sohini, and Sona.
If you use TactileNet in your research, please cite:
@misc{khan2025tactilenetbridgingaccessibilitygap,
title={TactileNet: Bridging the Accessibility Gap with AI-Generated Tactile Graphics for Individuals with Vision Impairment},
author={Adnan Khan and Alireza Choubineh and Mai A. Shaaban and Abbas Akkasi and Majid Komeili},
year={2025},
eprint={2504.04722},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2504.04722},
}