/* ScummVM - Graphic Adventure Engine * * ScummVM is the legal property of its developers, whose names * are too numerous to list here. Please refer to the COPYRIGHT * file distributed with this source distribution. * * This program is free software: you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation, either version 3 of the License, or * (at your option) any later version. * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * You should have received a copy of the GNU General Public License * along with this program. If not, see . * */ #include "common/random.h" #include "common/config-manager.h" #include "common/serializer.h" #include "common/memstream.h" #include "engines/nancy/nancy.h" #include "engines/nancy/sound.h" #include "engines/nancy/input.h" #include "engines/nancy/util.h" #include "engines/nancy/graphics.h" #include "engines/nancy/resource.h" #include "engines/nancy/action/conversation.h" #include "engines/nancy/state/scene.h" namespace Nancy { namespace Action { ConversationSound::ConversationSound() : RenderActionRecord(8), _noResponse(g_nancy->getGameType() <= kGameTypeNancy2 ? 10 : 20), _hasDrawnTextbox(false), _pickedResponse(-1) { _conditionalResponseCharacterID = _noResponse; _goodbyeResponseCharacterID = _noResponse; } ConversationSound::~ConversationSound() { if (NancySceneState.getActiveConversation() == this) { NancySceneState.setActiveConversation(nullptr); } } void ConversationSound::init() { RenderObject::init(); } void ConversationSound::readData(Common::SeekableReadStream &stream) { Common::Serializer ser(&stream, nullptr); ser.setVersion(g_nancy->getGameType()); if (ser.getVersion() >= kGameTypeNancy2) { _sound.readNormal(stream); } readCaptionText(stream); if (ser.getVersion() <= kGameTypeNancy1) { _sound.readNormal(stream); } _responseGenericSound.readNormal(stream); ser.skip(1); // RESPONSE_STARTUP_CLEAR_ALL, RESPONSE_STARTUP_KEEP_OLD; never used (tested up to nancy5) ser.syncAsByte(_conditionalResponseCharacterID); ser.syncAsByte(_goodbyeResponseCharacterID); ser.syncAsByte(_defaultNextScene); ser.syncAsByte(_popNextScene); _sceneChange.readData(stream, ser.getVersion() == kGameTypeVampire); ser.skip(0x32, kGameTypeVampire, kGameTypeNancy1); ser.skip(2, kGameTypeNancy2); uint16 numResponses = 0; ser.syncAsUint16LE(numResponses); _responses.resize(numResponses); for (uint i = 0; i < numResponses; ++i) { ResponseStruct &response = _responses[i]; response.conditionFlags.read(stream); readResponseText(stream, response); readFilename(stream, response.soundName); ser.syncAsByte(response.addRule); response.sceneChange.readData(stream, ser.getVersion() == kGameTypeVampire); ser.syncAsSint16LE(response.flagDesc.label); ser.syncAsByte(response.flagDesc.flag); ser.skip(0x32, kGameTypeVampire, kGameTypeNancy1); ser.skip(2, kGameTypeNancy2); } uint16 numSceneBranchStructs = stream.readUint16LE(); _sceneBranchStructs.resize(numSceneBranchStructs); for (uint i = 0; i < numSceneBranchStructs; ++i) { _sceneBranchStructs[i].conditions.read(stream); _sceneBranchStructs[i].sceneChange.readData(stream, g_nancy->getGameType() == kGameTypeVampire); ser.skip(0x32, kGameTypeVampire, kGameTypeNancy1); ser.skip(2, kGameTypeNancy2); } uint16 numFlagsStructs = stream.readUint16LE(); _flagsStructs.resize(numFlagsStructs); for (uint i = 0; i < numFlagsStructs; ++i) { FlagsStruct &flagsStruct = _flagsStructs[i]; flagsStruct.conditions.read(stream); flagsStruct.flagToSet.type = stream.readByte(); flagsStruct.flagToSet.flag.label = stream.readSint16LE(); flagsStruct.flagToSet.flag.flag = stream.readByte(); } } void ConversationSound::readTerseData(Common::SeekableReadStream &stream) { readFilename(stream, _sound.name); _sound.volume = stream.readUint16LE(); _sound.channelID = 12; // hardcoded _sound.numLoops = 1; _responseGenericSound.volume = _sound.volume; _responseGenericSound.numLoops = 1; _responseGenericSound.channelID = 13; // hardcoded readTerseCaptionText(stream); _conditionalResponseCharacterID = stream.readByte(); _goodbyeResponseCharacterID = stream.readByte(); _defaultNextScene = stream.readByte(); _sceneChange.sceneID = stream.readUint16LE(); _sceneChange.continueSceneSound = kContinueSceneSound; uint16 numResponses = stream.readUint16LE(); _responses.resize(numResponses); for (uint i = 0; i < numResponses; ++i) { ResponseStruct &response = _responses[i]; response.conditionFlags.read(stream); readTerseResponseText(stream, response); readFilename(stream, response.soundName); response.sceneChange.sceneID = stream.readUint16LE(); response.sceneChange.continueSceneSound = kContinueSceneSound; } // No scene branches uint16 numFlagsStructs = stream.readUint16LE(); _flagsStructs.resize(numFlagsStructs); for (uint i = 0; i < numFlagsStructs; ++i) { FlagsStruct &flagsStruct = _flagsStructs[i]; flagsStruct.conditions.read(stream); flagsStruct.flagToSet.type = stream.readByte(); flagsStruct.flagToSet.flag.label = stream.readSint16LE(); flagsStruct.flagToSet.flag.flag = stream.readByte(); } } void ConversationSound::readCaptionText(Common::SeekableReadStream &stream) { char *rawText = new char[1500]; stream.read(rawText, 1500); assembleTextLine(rawText, _text, 1500); delete[] rawText; } void ConversationSound::readResponseText(Common::SeekableReadStream &stream, ResponseStruct &response) { char *rawText = new char[400]; stream.read(rawText, 400); assembleTextLine(rawText, response.text, 400); delete[] rawText; } void ConversationSound::readTerseCaptionText(Common::SeekableReadStream &stream) { Common::String key; readFilename(stream, key); const CVTX *convo = (const CVTX *)g_nancy->getEngineData("CONVO"); assert(convo); _text = convo->texts[key]; } void ConversationSound::readTerseResponseText(Common::SeekableReadStream &stream, ResponseStruct &response) { Common::String key; readFilename(stream, key); const CVTX *convo = (const CVTX *)g_nancy->getEngineData("CONVO"); assert(convo); response.text = convo->texts[key]; } void ConversationSound::execute() { ConversationSound *activeConversation = NancySceneState.getActiveConversation(); if (activeConversation != this && activeConversation != nullptr) { if ( !activeConversation->_isDone || activeConversation->_defaultNextScene == kDefaultNextSceneEnabled || activeConversation->_pickedResponse != -1 ) { return; } else { // Chained videos, hide the previous one and start this activeConversation->setVisible(false); NancySceneState.setActiveConversation(this); } } switch (_state) { case kBegin: { init(); g_nancy->_sound->loadSound(_sound); if (!ConfMan.getBool("speech_mute") && ConfMan.getBool("character_speech")) { g_nancy->_sound->playSound(_sound); } // Remove held item and re-add it to inventory NancySceneState.setNoHeldItem(); // Move the mouse to the default position defined in CURS const Common::Point initialMousePos = g_nancy->_cursor->getPrimaryVideoInitialPos(); const Common::Point cursorHotspot = g_nancy->_cursor->getCurrentCursorHotspot(); Common::Point adjustedMousePos = g_nancy->_input->getInput().mousePos; adjustedMousePos.x -= cursorHotspot.x; adjustedMousePos.y -= cursorHotspot.y - 1; if (g_nancy->_cursor->getPrimaryVideoInactiveZone().bottom > adjustedMousePos.y) { g_nancy->_cursor->warpCursor(Common::Point(initialMousePos.x + cursorHotspot.x, initialMousePos.y + cursorHotspot.y)); g_nancy->_cursor->setCursorType(CursorManager::kNormalArrow); } _state = kRun; NancySceneState.setActiveConversation(this); // Do not draw first frame since video won't be loaded yet g_nancy->_graphics->suppressNextDraw(); if (g_nancy->getGameType() < kGameTypeNancy6) { // Do not fall through to give the execution one loop for event flag changes // This fixes TVD scene 750 break; } // However, nancy6 scene 1299 requires us to fall through in order to get the correct caption. // By that point Conversation scenes weren't the tangled mess they were in earlier games, // so hopefully this won't break anything } // fall through case kRun: if (!_hasDrawnTextbox) { _hasDrawnTextbox = true; auto *textboxData = GetEngineData(TBOX); assert(textboxData); NancySceneState.getTextbox().clear(); NancySceneState.getTextbox().setOverrideFont(textboxData->conversationFontID); if (ConfMan.getBool("subtitles")) { NancySceneState.getTextbox().addTextLine(_text); } Common::Array responsesToAdd; for (uint i = 0; i < _responses.size(); ++i) { auto &res = _responses[i]; if (res.conditionFlags.isSatisfied()) { int foundIndex = -1; for (uint j = 0; j < responsesToAdd.size(); ++j) { if (_responses[responsesToAdd[j]].text.compareToIgnoreCase(res.text) == 0) { foundIndex = j; break; } } switch(res.addRule) { case ResponseStruct::kAddIfNotFound: if (foundIndex == -1) { responsesToAdd.push_back(i); } break; case ResponseStruct::kRemoveAndAddToEnd: if (foundIndex != -1) { responsesToAdd.remove_at(foundIndex); } responsesToAdd.push_back(i); break; case ResponseStruct::kRemove: if (foundIndex != -1) { responsesToAdd.remove_at(foundIndex); } break; } } } uint numRegularResponses = _responses.size(); // Add responses when conditions have been satisfied if (_conditionalResponseCharacterID != _noResponse) { addConditionalDialogue(); } if (_goodbyeResponseCharacterID != _noResponse) { addGoodbye(); } // If conditionals/goodbyes added, make sure to send them to Textbox for (uint i = numRegularResponses; i < _responses.size(); ++i) { responsesToAdd.push_back(i); } for (uint i : responsesToAdd) { NancySceneState.getTextbox().addTextLine(_responses[i].text); _responses[i].isOnScreen = true; } } if (!g_nancy->_sound->isSoundPlaying(_sound) && isVideoDonePlaying()) { g_nancy->_sound->stopSound(_sound); bool hasResponses = false; for (auto &res : _responses) { if (res.isOnScreen) { hasResponses = true; break; } } if (!hasResponses) { // NPC has finished talking with no responses available, auto-advance to next scene _state = kActionTrigger; } else { // NPC has finished talking, we have responses for (uint i = 0; i < 30; ++i) { if (NancySceneState.getLogicCondition(i, g_nancy->_true)) { _pickedResponse = i; // Adjust to account for hidden responses for (uint j = 0; j < _responses.size(); ++j) { if (!_responses[j].isOnScreen) { ++_pickedResponse; } if ((int)j == _pickedResponse) { break; } } break; } } if (_pickedResponse != -1) { // Player has picked response, play sound file and change state _responseGenericSound.name = _responses[_pickedResponse].soundName; g_nancy->_sound->loadSound(_responseGenericSound); if (!ConfMan.getBool("speech_mute") && ConfMan.getBool("player_speech")) { g_nancy->_sound->playSound(_responseGenericSound); } _state = kActionTrigger; } } } break; case kActionTrigger: if (!g_nancy->_sound->isSoundPlaying(_responseGenericSound)) { // process flags structs for (auto &flags : _flagsStructs) { if (flags.conditions.isSatisfied()) { flags.flagToSet.set(); } } if (_pickedResponse != -1) { // Set response's event flag, if any NancySceneState.setEventFlag(_responses[_pickedResponse].flagDesc); } g_nancy->_sound->stopSound(_responseGenericSound); if (_pickedResponse != -1) { NancySceneState.changeScene(_responses[_pickedResponse].sceneChange); } else { for (uint i = 0; i < _sceneBranchStructs.size(); ++i) { if (_sceneBranchStructs[i].conditions.isSatisfied()) { NancySceneState.changeScene(_sceneBranchStructs[i].sceneChange); break; } } if (_defaultNextScene == kDefaultNextSceneEnabled) { NancySceneState.changeScene(_sceneChange); } else if (_popNextScene == kPopNextScene) { // Exit dialogue NancySceneState.popScene(); } } finishExecution(); } break; } } void ConversationSound::addConditionalDialogue() { for (const auto &res : g_nancy->getStaticData().conditionalDialogue[_conditionalResponseCharacterID]) { bool isSatisfied = true; for (const auto &cond : res.conditions) { switch (cond.type) { case (byte)StaticDataConditionType::kEvent : if (!NancySceneState.getEventFlag(cond.label, cond.flag)) { isSatisfied = false; } break; case (byte)StaticDataConditionType::kInventory : if (NancySceneState.hasItem(cond.label) != cond.flag) { isSatisfied = false; } break; case (byte)StaticDataConditionType::kDifficulty : if ( (NancySceneState.getDifficulty() != cond.label && cond.flag != 0) || (NancySceneState.getDifficulty() == cond.label && cond.flag == 0) ) { isSatisfied = false; } break; } if (!isSatisfied) { break; } } if (isSatisfied) { _responses.push_back(ResponseStruct()); ResponseStruct &newResponse = _responses.back(); newResponse.soundName = res.soundID; if (g_nancy->getGameType() <= kGameTypeNancy5) { // String is also inside nancy.dat newResponse.text = g_nancy->getStaticData().conditionalDialogueTexts[res.textID]; } else { // String is inside the CVTX chunk in the CONVO file. Sound ID doubles as string key const CVTX *convo = (const CVTX *)g_nancy->getEngineData("CONVO"); assert(convo); newResponse.text = convo->texts[res.soundID]; } newResponse.sceneChange.sceneID = res.sceneID; newResponse.sceneChange.continueSceneSound = kContinueSceneSound; newResponse.sceneChange.listenerFrontVector.set(0, 0, 1); // Check if the response is a repeat. This can happen when multiple condition combinations // trigger the same response. for (uint i = 0; i < _responses.size() - 1; ++i) { if ( _responses[i].soundName == newResponse.soundName && _responses[i].text == newResponse.text && _responses[i].sceneChange.sceneID == newResponse.sceneChange.sceneID) { _responses.pop_back(); break; } } } } } void ConversationSound::addGoodbye() { auto &res = g_nancy->getStaticData().goodbyes[_goodbyeResponseCharacterID]; _responses.push_back(ResponseStruct()); ResponseStruct &newResponse = _responses.back(); newResponse.soundName = res.soundID; if (g_nancy->getGameType() <= kGameTypeNancy5) { // String is also inside nancy.dat newResponse.text = g_nancy->getStaticData().goodbyeTexts[_goodbyeResponseCharacterID]; } else { // String is inside the CVTX chunk in the CONVO file. Sound ID doubles as string key const CVTX *convo = (const CVTX *)g_nancy->getEngineData("CONVO"); assert(convo); newResponse.text = convo->texts[res.soundID]; } // Evaluate conditions to pick from the collection of replies uint sceneChangeID = 0; for (uint i = 0; i < res.sceneChanges.size(); ++i) { const GoodbyeSceneChange &sc = res.sceneChanges[i]; if (sc.conditions.size() == 0) { // No conditions, default choice sceneChangeID = i; break; } else { bool isSatisfied = true; for (const auto &cond : sc.conditions) { switch (cond.type) { case (byte)StaticDataConditionType::kEvent : if (!NancySceneState.getEventFlag(cond.label, cond.flag)) { isSatisfied = false; } break; case (byte)StaticDataConditionType::kInventory : if (NancySceneState.hasItem(cond.label) != cond.flag) { isSatisfied = false; } break; case (byte)StaticDataConditionType::kDifficulty : if ( (NancySceneState.getDifficulty() != cond.label && cond.flag != 0) || (NancySceneState.getDifficulty() == cond.label && cond.flag == 0) ) { isSatisfied = false; } break; } if (!isSatisfied) { break; } } if (isSatisfied) { sceneChangeID = i; break; } } } const GoodbyeSceneChange &sceneChange = res.sceneChanges[sceneChangeID]; // The reply from the character is picked randomly newResponse.sceneChange.sceneID = sceneChange.sceneIDs[g_nancy->_randomSource->getRandomNumber(sceneChange.sceneIDs.size() - 1)]; newResponse.sceneChange.continueSceneSound = kContinueSceneSound; newResponse.sceneChange.listenerFrontVector.set(0, 0, 1); // Set an event flag if applicable // Assumes flagToSet is an event flag NancySceneState.setEventFlag(sceneChange.flagToSet.label, sceneChange.flagToSet.flag); } void ConversationSound::ConversationFlag::read(Common::SeekableReadStream &stream) { type = stream.readByte(); flag.label = stream.readSint16LE(); flag.flag = stream.readByte(); orFlag = stream.readByte(); } bool ConversationSound::ConversationFlag::isSatisfied() const { switch (type) { case kFlagEvent: return NancySceneState.getEventFlag(flag); case kFlagInventory: return NancySceneState.hasItem(flag.label) == flag.flag; default: return false; } } void ConversationSound::ConversationFlag::set() const { switch (type) { case kFlagEvent: NancySceneState.setEventFlag(flag); break; case kFlagInventory: if (flag.flag == g_nancy->_true) { NancySceneState.addItemToInventory(flag.label); } else { NancySceneState.removeItemFromInventory(flag.label); } break; default: break; } } void ConversationSound::ConversationFlags::read(Common::SeekableReadStream &stream) { uint16 numFlags = stream.readUint16LE(); conditionFlags.resize(numFlags); for (uint i = 0; i < numFlags; ++i) { conditionFlags[i].read(stream); } } bool ConversationSound::ConversationFlags::isSatisfied() const { Common::Array conditionsMet(conditionFlags.size(), false); for (uint i = 0; i < conditionFlags.size(); ++i) { if (conditionFlags[i].isSatisfied()) { conditionsMet[i] = true; } } for (uint i = 0; i < conditionsMet.size(); ++i) { if (conditionFlags[i].orFlag) { bool foundSatisfied = false; for (uint j = 0; j < conditionFlags.size(); ++j) { if (conditionsMet[j]) { foundSatisfied = true; break; } // Found end of orFlag chain if (!conditionFlags[j].orFlag) { break; } } if (foundSatisfied) { for (; i < conditionsMet.size(); ++i) { conditionsMet[i] = true; if (!conditionFlags[i].orFlag) { // End of orFlag chain break; } } } } } for (uint i = 0; i < conditionsMet.size(); ++i) { if (conditionsMet[i] == false) { return false; } } return true; } void ConversationVideo::init() { if (!_decoder.loadFile(Common::Path(_videoName + ".avf"))) { error("Couldn't load video file %s", _videoName.c_str()); } _decoder.seekToFrame(_firstFrame); if (!_paletteName.empty()) { GraphicsManager::loadSurfacePalette(_drawSurface, _paletteName); setTransparent(true); } ConversationSound::init(); registerGraphics(); } void ConversationVideo::readData(Common::SeekableReadStream &stream) { Common::Serializer ser(&stream, nullptr); ser.setVersion(g_nancy->getGameType()); readFilename(stream, _videoName); readFilename(ser, _paletteName, kGameTypeVampire, kGameTypeVampire); ser.skip(2, kGameTypeVampire, kGameTypeNancy1); ser.syncAsUint16LE(_videoFormat); ser.skip(3); // Quality ser.skip(4, kGameTypeVampire, kGameTypeVampire); // paletteStart, paletteSize ser.syncAsUint16LE(_firstFrame); ser.syncAsUint16LE(_lastFrame); ser.skip(8, kGameTypeVampire, kGameTypeNancy1); ser.skip(6, kGameTypeNancy2); ser.skip(0x10); // Bounds readRect(stream, _screenPosition); ConversationSound::readData(stream); } void ConversationVideo::updateGraphics() { if (!_decoder.isVideoLoaded()) { return; } if (!_decoder.isPlaying()) { _decoder.start(); } if (_decoder.getCurFrame() == _lastFrame) { _decoder.pauseVideo(true); } if (_decoder.needsUpdate()) { GraphicsManager::copyToManaged(*_decoder.decodeNextFrame(), _drawSurface, _videoFormat == kSmallVideoFormat); _needsRedraw = true; } RenderObject::updateGraphics(); } void ConversationVideo::onPause(bool pause) { _decoder.pauseVideo(pause); RenderActionRecord::onPause(pause); } bool ConversationVideo::isVideoDonePlaying() { return _decoder.endOfVideo() || _decoder.getCurFrame() == _lastFrame; } Common::String ConversationVideo::getRecordTypeName() const { if (g_nancy->getGameType() <= kGameTypeNancy1) { return "PlayPrimaryVideo"; } else { return "ConversationVideo"; } } class ConversationCelLoader : public DeferredLoader { public: ConversationCelLoader(ConversationCel &owner) : _owner(owner) {} private: bool loadInner() override; ConversationCel &_owner; }; bool ConversationCelLoader::loadInner() { for (uint i = _owner._curFrame; i < _owner._celNames[0].size(); ++i) { for (uint j = 0; j < _owner._celRObjects.size(); ++j) { if (!_owner._celCache.contains(_owner._celNames[j][i])) { _owner.loadCel(_owner._celNames[j][i], _owner._treeNames[j]); return false; } } } return true; } ConversationCel::~ConversationCel() { // Make sure there isn't a single-frame gap between conversation scenes where // the character is invisible g_nancy->_graphics->suppressNextDraw(); } void ConversationCel::init() { _curFrame = _firstFrame; _nextFrameTime = g_nancy->getTotalPlayTime(); ConversationSound::init(); _loaderPtr.reset(new ConversationCelLoader(*this)); auto castedPtr = _loaderPtr.dynamicCast(); g_nancy->addDeferredLoader(castedPtr); for (uint i = 0; i < _treeNames.size(); ++i) { if (_treeNames[i].size()) { _celRObjects.push_back(RenderedCel()); } else break; } registerGraphics(); } void ConversationCel::registerGraphics() { for (uint i = 0; i < _celRObjects.size(); ++i) { _celRObjects[i]._z = 9 + _drawingOrder[i]; _celRObjects[i].setVisible(true); _celRObjects[i].setTransparent(true); _celRObjects[i].registerGraphics(); } RenderActionRecord::registerGraphics(); } void ConversationCel::updateGraphics() { uint32 currentTime = g_nancy->getTotalPlayTime(); if (_state == kRun && currentTime > _nextFrameTime && _curFrame < MIN(_lastFrame + 1, _celNames[0].size())) { for (uint i = 0; i < _celRObjects.size(); ++i) { Cel &cel = loadCel(_celNames[i][_curFrame], _treeNames[i]); if (_overrideTreeRects[i] == kCelOverrideTreeRectsOn) { _celRObjects[i]._drawSurface.create(cel.surf, _overrideRectSrcs[i]); _celRObjects[i].moveTo(_overrideRectDests[i]); } else { _celRObjects[i]._drawSurface.create(cel.surf, cel.src); _celRObjects[i].moveTo(cel.dest); } } _nextFrameTime += _frameTime; ++_curFrame; } } void ConversationCel::readData(Common::SeekableReadStream &stream) { Common::String xsheetName; readFilename(stream, xsheetName); readFilenameArray(stream, _treeNames, 4); readXSheet(stream, xsheetName); // Continue reading the AR stream // Something related to quality stream.skip(3); _firstFrame = stream.readUint16LE(); _lastFrame = stream.readUint16LE(); stream.skip(6); _drawingOrder.resize(4); for (uint i = 0; i < 4; ++i) { _drawingOrder[i] = stream.readByte(); } _overrideTreeRects.resize(4); for (uint i = 0; i < 4; ++i) { _overrideTreeRects[i] = stream.readByte(); } readRectArray(stream, _overrideRectSrcs, 4); readRectArray(stream, _overrideRectDests, 4); ConversationSound::readData(stream); } void ConversationCel::readXSheet(Common::SeekableReadStream &stream, const Common::String &xsheetName) { Common::SeekableReadStream *xsheet = SearchMan.createReadStreamForMember(Common::Path(xsheetName)); // Read the xsheet and load all images into the arrays // Completely unoptimized, the original engine uses a buffer xsheet->seek(0); Common::String signature = xsheet->readString('\0', 18); if (signature != "XSHEET WayneSikes") { warning("XSHEET signature doesn't match!"); return; } xsheet->seek(0x22); uint numFrames = xsheet->readUint16LE(); xsheet->skip(2); _frameTime = xsheet->readUint16LE(); xsheet->skip(2); _celNames.resize(4, Common::Array(numFrames)); for (uint i = 0; i < numFrames; ++i) { for (uint j = 0; j < _celNames.size(); ++j) { readFilename(*xsheet, _celNames[j][i]); } // 4 unknown values xsheet->skip(8); } delete xsheet; } bool ConversationCel::isVideoDonePlaying() { return _curFrame >= MIN(_lastFrame, _celNames[0].size()) && _nextFrameTime <= g_nancy->getTotalPlayTime(); } ConversationCel::Cel &ConversationCel::loadCel(const Common::Path &name, const Common::String &treeName) { // Assumes head and body cels will be named differently if (_celCache.contains(name)) { return _celCache[name]; } Cel &newCel = _celCache.getOrCreateVal(name); g_nancy->_resource->loadImage(name, newCel.surf, treeName, &newCel.src, &newCel.dest); return _celCache[name]; } void ConversationSoundTerse::readData(Common::SeekableReadStream &stream) { readTerseData(stream); } void ConversationCelTerse::readData(Common::SeekableReadStream &stream) { Common::String xsheetName; readFilename(stream, xsheetName); readFilenameArray(stream, _treeNames, 2); // Only 2 readXSheet(stream, xsheetName); _lastFrame = stream.readUint16LE(); _drawingOrder = { 1, 0, 2, 3 }; _overrideTreeRects.resize(4, kCelOverrideTreeRectsOff); readTerseData(stream); } } // End of namespace Action } // End of namespace Nancy