Proposal
Problem
Classification on with tabular using deep neural networks
Possible solutions
First Approach
Contrastive Learning for Tabular Data Pretraining
Using a TabContrast approach:
- Generate positive pairs by applying augmentations to your tabular data:
- Implement an encoder architecture
- Use NT-Xent loss:
- Maximize similarity between positive pairs
- Minimize similarity between negative pairs (other samples in batch)
I mean I could use something else, but I think this method allows for more artifically generated training data that could be beneficial.
Maybe ReConTab + TabContrast but that may be too complex
Transforming Table Rows into Images
- Use: REFINED, etc to convert to image data
- Use pre-trained CNN to extract features.
Crazy to use contrast learning on this phase?
Merging the Two Information Sources
Using an attention-based fusion approach:
- For tabular data: Extract the lower-dimensional embeddings form our encoder.
- For images: use our CNN to extract the lower-dimensional representation of the image.
- Fusion:
- Project embeddings on the same dimension (this might require another head?)
- Implement cross-attention (… I don’t know how this works)
- Pass through a final MLP for classification
Second Approach
Given a tabular dataset use multimodal learning where the two modalities are as follows:
- Generate image data same as before, using REFINED, etc.
- Tabular data: raw dataset.
Use directly the approach presented on Best of Both Worlds: Multimodal Contrastive Learning with Tabular and Imaging Data