Sean Moran Sean Moran
Research

Topics I work on.

A short tour of the threads that run through my published work — what I find interesting, what I keep returning to, and the older papers behind the more recent ones.

Retrieval and representation learning

My PhD at EdinburghNLP with Victor Lavrenko was on hash codes for large-scale image retrieval — the line of work that became Variable Bit Quantisation (ACL 2013), Neighbourhood-Preserving Quantisation (SIGIR 2013), Graph-Regularised Hashing (ECIR 2015), and a 2019 monograph on the field. Sparse Kernel Learning for Image Annotation (ICMR 2014) won Best Student Paper. The throughline is the same question I still find interesting: how do you build representations that let you find the right thing in a huge collection, fast.

Some of my recent essays — on the bits-over-random metric, on why some RAG queries can't be solved by vector search alone — are extensions of that thinking into the LLM era.

Image enhancement and small interpretable components

A second strand is photo-quality image enhancement. DeepLPF (CVPR 2020) introduced learnable parametric filters, CURL (ICPR 2020) introduced neural curve layers, and SIDGAN (ECCV 2020) introduced a synthetic-data pipeline for low-light video. The shared idea: small, interpretable modules that get most of the benefit of much heavier networks.

Applied AI in regulated environments

For the last several years I have been part of JPMorgan's AI Research group, working on the deployment of generative AI in one of the most regulated computing environments out there. Published work from that period — SpamT5 (FinLLM @ IJCAI 2023), CodeQUEST (ISSREW 2025), API-Miner, Senatus, Ledgit, DeepClean — is mostly about making large models do real, audited work: code-quality evaluation, anomaly detection, federated training on sensitive data, machine unlearning. The patent portfolio is the residue of that effort.

How I tend to work


Full list of papers and patents on the home page.