Looking at the Glk specification 0.7.5, it appears that developers could, in the near-term, bridge Glk and the Bot Framework for text-based and hypertext-based scenarios. Many Bot Framework channels are text-based or hypertext-based and, after some software development, existing, current, and upcoming interactive fiction games could work across all of these channels.
In the longer-term, developers could use the Bot Framework to enable: (a) multi-channel, (b) smart-speaker, and (c) video-calling-based games. By multi-channel games, I mean games where players could pause play from one channel, e.g., a smartphone, and resume play on another channel, e.g., using a PC or a smart speaker – effectively playing across devices. This would be possible with server-side and/or cloud-based software engines. By smart-speaker scenarios, I mean uses of the Bot Framework to deliver user experiences on Siri, Alexa, Google Assistant, Cortana, et al. By video-calling-based games, I mean uses of the Bot Framework to deliver audiovisual user experiences across channels like WebRTC, Zoom, WebEx, or Skype.
In these regards, exciting possibilities with respect to the futures of the Glk specification include enabling speech-recognition and natural-language-understanding scenarios.
With respect to speech recognition, there exists a standard: the Speech Recognition Grammar Specification (SRGS). Speech recognition grammars define a subset of all natural-language utterances that a machine might expect. Dynamic grammars would be a relevant feature for interactive fiction games. As contexts change, so too could the grammars describing the instantaneously expected natural language. As contexts change, e.g., as players move between rooms, perhaps Glk (version next) could manage dynamic SRGS grammars – in a platform-independent manner – enabling speech recognition and conversational user interfaces for interactive fiction games.
With respect to natural-language understanding, there are also, more recently, approaches such as LUIS (documentation). Services like LUIS can process across paraphrases. So, utterances like “pick up the lamp”, “get the lamp”, “grab the lamp”, and so forth, could all be processed by services like LUIS which return the same data structures across paraphrases.
It appears that (1) developers can bridge Glk and the Bot Framework, delivering text-based and hypertext-based channels, in the near term, and that (2) towards (a) multi-channel, (b) smart-speaker, and (c) video-calling-based games, speech-recognition (SRGS) and natural-language understanding (LUIS) capabilities are possible on the horizon.