Peter Vamplew
@amp1874.bsky.social
210 followers 220 following 220 posts
Professor in IT @ Federation Uni. Multi-objective reinforcement learning. Human-aligned AI. Best known for the f*cking mailing list paper. Jambo & Bengals fan. https://t.co/UNoOrbGApz
Posts Media Videos Starter Packs
amp1874.bsky.social
Dear Benjamin,

Congrats on being the fastest scammy conference organiser of all time. Inviting me to a conference unrelated to the topic of my paper is highly questionable, but you did it within a day of publication so at least you’re fast.

Kindly remove me from your mailing list.

Regards,
Peter
amp1874.bsky.social
It examines the components of effective human apologies, and analyses how and how well these have been implemented in prior apologetic AI systems. Haddie has done a fantastic job here - this is the most thorough and in-depth student publication that I've been fortunate enough to be involved in.
amp1874.bsky.social
Computer says sorry?

After months in copy-editing hell, Haddie Harland's review of AI apology research is now available: link.springer.com/article/10.1...

This is a must read for anyone interested in how AI systems can effectively and appropriately use apologies to facilitate human interaction 1/2
AI apology: a critical review of apology in AI systems - Artificial Intelligence Review
Apologies are a powerful tool used in human-human interactions to provide affective support, regulate social processes, and exchange information following a trust violation. The emerging field of AI apology investigates the use of apologies by artificially intelligent systems, with recent research suggesting how this tool may provide similar value in human-machine interactions. Until recently, contributions to this area were sparse, and these works have yet to be synthesised into a cohesive body of knowledge. This article provides the first synthesis and critical analysis of the state of AI apology research, focusing on studies published between 2020 and 2023. We derive a framework of attributes to describe five core elements of apology: outcome, interaction, offence, recipient, and offender. With this framework as the basis for our critique, we show how apologies can be used to recover from misalignment in human-AI interactions, and examine trends and inconsistencies within the field. Among the observations, we outline the importance of curating a human-aligned and cross-disciplinary perspective in this research, with consideration for improved system capabilities and long-term outcomes.
link.springer.com
amp1874.bsky.social
Have it also make the initial suggestion of a dish, otherwise there's a risk you might accidentally choose something which you do have the ingredients for.
amp1874.bsky.social
Decisions, decisions. Which should I read first?
Reposted by Peter Vamplew
cslg-bot.bsky.social
Dsouza, Ofosu, Amaogu, Pigeon, Boudreault, Maghoul, Moreno-Cruz, Leonenko: BoreaRL: A Multi-Objective Reinforcement Learning Environment for Climate-Adaptive Boreal Forest Management https://arxiv.org/abs/2509.19846 https://arxiv.org/pdf/2509.19846 https://arxiv.org/html/2509.19846
Reposted by Peter Vamplew
arxiv-cs-cl.bsky.social
Lingxiao Kong, Cong Yang, Oya Deniz Beyan, Zeyd Boukhers
Multi-Objective Reinforcement Learning for Large Language Model Optimization: Visionary Perspective
https://arxiv.org/abs/2509.21613
Reposted by Peter Vamplew
tmiller-uq.bsky.social
The deadline for my postdoc on scalable clinical decision support is closing in 1 week: 4 October (Australian Eastern standard Time). Please share with anyone that you think would be interested
tmiller-uq.bsky.social
I'm hiring again! Please share. I'm recruiting a postdoc research fellow in human-centred AI for scalable decision support. Join us to investigate how to balance scalability and human control in medical decision support. Closing date: 4 October (AEST).
uqtmiller.github.io/recruitment/
Recruitment
uqtmiller.github.io
amp1874.bsky.social
@tresvillain.bsky.social Andrew, for consistency you need to change your name to Trendvillian :-)
amp1874.bsky.social
Dear authors who I shall not name. Thank you for citing my work. But I'm not sure that a paper dating from 1995 should be cited in the context of a paragraph which begins "Recent trends show...."
amp1874.bsky.social
It might be skewed by the topics I search, but I find that almost every response I get contains at least one statement which is clearly wrong. So I just skip past them these days and go to the search results. It's not just Google, I've found Copilot Pro to be just as bad.
amp1874.bsky.social
Why would you love them? My experience is that they are blatantly incorrect about 90% of the time.
Reposted by Peter Vamplew
upolehsan.bsky.social
⚠️ The #CHI2026 paper I submitted? It almost didn't exist. That's the BTS part academics never post. So I will…to normalize what I call unglamorous persistence.

This summer was one of my hardest, mentally. 🌥️ Between ...
1/n
amp1874.bsky.social
Hey! There's finally someone else in Australia doing research in multi-objective reinforcement learning. @marcusgal.bsky.social arxiv.org/pdf/2509.14816
arxiv.org
amp1874.bsky.social
Interesting. We've noticed that changes in greedy policy can cause interference in the vector values learned by a multi-objective RL agent which hurts learning, but hadn't measured how frequently that happens. This paper suggests it might be a bigger issue than we thought.
Reposted by Peter Vamplew
amp1874.bsky.social
[42] 42. Liu, H., et al.: Global and local structure preserving network for 3D human pose estimation. IEEE Trans. Image Process. 30, 1158–1171 (2021)
amp1874.bsky.social
[41] Ren, Z., et al.: Structure-aware generation network for anatomically plausible brain image synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9440–9449 (2019)
amp1874.bsky.social
[38] Sohl-Dickstein, J., et al.: The energy diffusion model: training energy-based models in diffusion time. arXiv preprint arXiv:2006.11239 (2020)
Note: Not only is there no paper with this title, the arXiv link given is for the previous paper in the reference list [37].
amp1874.bsky.social
[3] Karras, T., Laine, S., Aila, T.: StyleGAN2: improved styleGAN for realistic image synthesis. In: Proceedings of the IEEE/CVF International Conference on ComputerVision (ICCV) (2019)
amp1874.bsky.social
Given my doubts about this reference, I have checked the other references in this chapter using the same approach. While the majority of the references are genuine, my investigations indicate that the following references also do not correspond to actual publications:
amp1874.bsky.social
I can only conclude that this paper does not exist, and was either invented by Fahim and Maji, or (more likely) indicates that generative AI was used inappropriately in the production of this chapter.
amp1874.bsky.social
Searching Google Scholar, the only match is the citation in Fahim and Maji’s chapter. Similarly doing a Google search for that exact title in quotes, the only matches are to Fahim and Maji’s chapter.