| 1 | /* |
| 2 | * Copyright (C) 2008, 2009 Apple Inc. All rights reserved. |
| 3 | * Copyright (C) 2008 Cameron Zwarich <cwzwarich@uwaterloo.ca> |
| 4 | * |
| 5 | * Redistribution and use in source and binary forms, with or without |
| 6 | * modification, are permitted provided that the following conditions |
| 7 | * are met: |
| 8 | * |
| 9 | * 1. Redistributions of source code must retain the above copyright |
| 10 | * notice, this list of conditions and the following disclaimer. |
| 11 | * 2. Redistributions in binary form must reproduce the above copyright |
| 12 | * notice, this list of conditions and the following disclaimer in the |
| 13 | * documentation and/or other materials provided with the distribution. |
| 14 | * 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of |
| 15 | * its contributors may be used to endorse or promote products derived |
| 16 | * from this software without specific prior written permission. |
| 17 | * |
| 18 | * THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY |
| 19 | * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED |
| 20 | * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| 21 | * DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY |
| 22 | * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES |
| 23 | * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; |
| 24 | * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND |
| 25 | * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| 26 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
| 27 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| 28 | */ |
| 29 | |
| 30 | #include "config.h" |
| 31 | #include "BytecodeGenerator.h" |
| 32 | |
| 33 | #include "BatchedTransitionOptimizer.h" |
| 34 | #include "PrototypeFunction.h" |
| 35 | #include "JSFunction.h" |
| 36 | #include "Interpreter.h" |
| 37 | #include "UString.h" |
| 38 | |
| 39 | using namespace std; |
| 40 | |
| 41 | namespace JSC { |
| 42 | |
| 43 | /* |
| 44 | The layout of a register frame looks like this: |
| 45 | |
| 46 | For |
| 47 | |
| 48 | function f(x, y) { |
| 49 | var v1; |
| 50 | function g() { } |
| 51 | var v2; |
| 52 | return (x) * (y); |
| 53 | } |
| 54 | |
| 55 | assuming (x) and (y) generated temporaries t1 and t2, you would have |
| 56 | |
| 57 | ------------------------------------ |
| 58 | | x | y | g | v2 | v1 | t1 | t2 | <-- value held |
| 59 | ------------------------------------ |
| 60 | | -5 | -4 | -3 | -2 | -1 | +0 | +1 | <-- register index |
| 61 | ------------------------------------ |
| 62 | | params->|<-locals | temps-> |
| 63 | |
| 64 | Because temporary registers are allocated in a stack-like fashion, we |
| 65 | can reclaim them with a simple popping algorithm. The same goes for labels. |
| 66 | (We never reclaim parameter or local registers, because parameters and |
| 67 | locals are DontDelete.) |
| 68 | |
| 69 | The register layout before a function call looks like this: |
| 70 | |
| 71 | For |
| 72 | |
| 73 | function f(x, y) |
| 74 | { |
| 75 | } |
| 76 | |
| 77 | f(1); |
| 78 | |
| 79 | > <------------------------------ |
| 80 | < > reserved: call frame | 1 | <-- value held |
| 81 | > >snip< <------------------------------ |
| 82 | < > +0 | +1 | +2 | +3 | +4 | +5 | <-- register index |
| 83 | > <------------------------------ |
| 84 | | params->|<-locals | temps-> |
| 85 | |
| 86 | The call instruction fills in the "call frame" registers. It also pads |
| 87 | missing arguments at the end of the call: |
| 88 | |
| 89 | > <----------------------------------- |
| 90 | < > reserved: call frame | 1 | ? | <-- value held ("?" stands for "undefined") |
| 91 | > >snip< <----------------------------------- |
| 92 | < > +0 | +1 | +2 | +3 | +4 | +5 | +6 | <-- register index |
| 93 | > <----------------------------------- |
| 94 | | params->|<-locals | temps-> |
| 95 | |
| 96 | After filling in missing arguments, the call instruction sets up the new |
| 97 | stack frame to overlap the end of the old stack frame: |
| 98 | |
| 99 | |----------------------------------> < |
| 100 | | reserved: call frame | 1 | ? < > <-- value held ("?" stands for "undefined") |
| 101 | |----------------------------------> >snip< < |
| 102 | | -7 | -6 | -5 | -4 | -3 | -2 | -1 < > <-- register index |
| 103 | |----------------------------------> < |
| 104 | | | params->|<-locals | temps-> |
| 105 | |
| 106 | That way, arguments are "copied" into the callee's stack frame for free. |
| 107 | |
| 108 | If the caller supplies too many arguments, this trick doesn't work. The |
| 109 | extra arguments protrude into space reserved for locals and temporaries. |
| 110 | In that case, the call instruction makes a real copy of the call frame header, |
| 111 | along with just the arguments expected by the callee, leaving the original |
| 112 | call frame header and arguments behind. (The call instruction can't just discard |
| 113 | extra arguments, because the "arguments" object may access them later.) |
| 114 | This copying strategy ensures that all named values will be at the indices |
| 115 | expected by the callee. |
| 116 | */ |
| 117 | |
| 118 | #ifndef NDEBUG |
| 119 | static bool s_dumpsGeneratedCode = false; |
| 120 | #endif |
| 121 | |
| 122 | void BytecodeGenerator::setDumpsGeneratedCode(bool dumpsGeneratedCode) |
| 123 | { |
| 124 | #ifndef NDEBUG |
| 125 | s_dumpsGeneratedCode = dumpsGeneratedCode; |
| 126 | #else |
| 127 | UNUSED_PARAM(dumpsGeneratedCode); |
| 128 | #endif |
| 129 | } |
| 130 | |
| 131 | bool BytecodeGenerator::dumpsGeneratedCode() |
| 132 | { |
| 133 | #ifndef NDEBUG |
| 134 | return s_dumpsGeneratedCode; |
| 135 | #else |
| 136 | return false; |
| 137 | #endif |
| 138 | } |
| 139 | |
| 140 | void BytecodeGenerator::generate() |
| 141 | { |
| 142 | m_codeBlock->setThisRegister(m_thisRegister.index()); |
| 143 | |
| 144 | m_scopeNode->emitBytecode(*this); |
| 145 | |
| 146 | #ifndef NDEBUG |
| 147 | m_codeBlock->setInstructionCount(m_codeBlock->instructions().size()); |
| 148 | |
| 149 | if (s_dumpsGeneratedCode) |
| 150 | m_codeBlock->dump(m_scopeChain->globalObject()->globalExec()); |
| 151 | #endif |
| 152 | |
| 153 | if ((m_codeType == FunctionCode && !m_codeBlock->needsFullScopeChain() && !m_codeBlock->usesArguments()) || m_codeType == EvalCode) |
| 154 | symbolTable().clear(); |
| 155 | |
| 156 | m_codeBlock->setIsNumericCompareFunction(instructions() == m_globalData->numericCompareFunction(m_scopeChain->globalObject()->globalExec())); |
| 157 | |
| 158 | #if !ENABLE(OPCODE_SAMPLING) |
| 159 | if (!m_regeneratingForExceptionInfo && (m_codeType == FunctionCode || m_codeType == EvalCode)) |
| 160 | m_codeBlock->clearExceptionInfo(); |
| 161 | #endif |
| 162 | |
| 163 | m_codeBlock->shrinkToFit(); |
| 164 | } |
| 165 | |
| 166 | bool BytecodeGenerator::addVar(const Identifier& ident, bool isConstant, RegisterID*& r0) |
| 167 | { |
| 168 | int index = m_calleeRegisters.size(); |
| 169 | SymbolTableEntry newEntry(index, isConstant ? ReadOnly : 0); |
| 170 | pair<SymbolTable::iterator, bool> result = symbolTable().add(key: ident.ustring().rep(), mapped: newEntry); |
| 171 | |
| 172 | if (!result.second) { |
| 173 | r0 = ®isterFor(index: result.first->second.getIndex()); |
| 174 | return false; |
| 175 | } |
| 176 | |
| 177 | ++m_codeBlock->m_numVars; |
| 178 | r0 = newRegister(); |
| 179 | return true; |
| 180 | } |
| 181 | |
| 182 | bool BytecodeGenerator::addGlobalVar(const Identifier& ident, bool isConstant, RegisterID*& r0) |
| 183 | { |
| 184 | int index = m_nextGlobalIndex; |
| 185 | SymbolTableEntry newEntry(index, isConstant ? ReadOnly : 0); |
| 186 | pair<SymbolTable::iterator, bool> result = symbolTable().add(key: ident.ustring().rep(), mapped: newEntry); |
| 187 | |
| 188 | if (!result.second) |
| 189 | index = result.first->second.getIndex(); |
| 190 | else { |
| 191 | --m_nextGlobalIndex; |
| 192 | m_globals.append(value: index + m_globalVarStorageOffset); |
| 193 | } |
| 194 | |
| 195 | r0 = ®isterFor(index); |
| 196 | return result.second; |
| 197 | } |
| 198 | |
| 199 | void BytecodeGenerator::preserveLastVar() |
| 200 | { |
| 201 | if ((m_firstConstantIndex = m_calleeRegisters.size()) != 0) |
| 202 | m_lastVar = &m_calleeRegisters.last(); |
| 203 | } |
| 204 | |
| 205 | BytecodeGenerator::BytecodeGenerator(ProgramNode* programNode, const Debugger* debugger, const ScopeChain& scopeChain, SymbolTable* symbolTable, ProgramCodeBlock* codeBlock) |
| 206 | : m_shouldEmitDebugHooks(!!debugger) |
| 207 | , m_shouldEmitProfileHooks(scopeChain.globalObject()->supportsProfiling()) |
| 208 | , m_scopeChain(&scopeChain) |
| 209 | , m_symbolTable(symbolTable) |
| 210 | , m_scopeNode(programNode) |
| 211 | , m_codeBlock(codeBlock) |
| 212 | , m_thisRegister(RegisterFile::ProgramCodeThisRegister) |
| 213 | , m_finallyDepth(0) |
| 214 | , m_dynamicScopeDepth(0) |
| 215 | , m_baseScopeDepth(0) |
| 216 | , m_codeType(GlobalCode) |
| 217 | , m_nextGlobalIndex(-1) |
| 218 | , m_nextConstantOffset(0) |
| 219 | , m_globalConstantIndex(0) |
| 220 | , m_globalData(&scopeChain.globalObject()->globalExec()->globalData()) |
| 221 | , m_lastOpcodeID(op_end) |
| 222 | , m_emitNodeDepth(0) |
| 223 | , m_regeneratingForExceptionInfo(false) |
| 224 | , m_codeBlockBeingRegeneratedFrom(0) |
| 225 | { |
| 226 | if (m_shouldEmitDebugHooks) |
| 227 | m_codeBlock->setNeedsFullScopeChain(true); |
| 228 | |
| 229 | emitOpcode(op_enter); |
| 230 | codeBlock->setGlobalData(m_globalData); |
| 231 | |
| 232 | // FIXME: Move code that modifies the global object to Interpreter::execute. |
| 233 | |
| 234 | m_codeBlock->m_numParameters = 1; // Allocate space for "this" |
| 235 | |
| 236 | JSGlobalObject* globalObject = scopeChain.globalObject(); |
| 237 | ExecState* exec = globalObject->globalExec(); |
| 238 | RegisterFile* registerFile = &exec->globalData().interpreter->registerFile(); |
| 239 | |
| 240 | // Shift register indexes in generated code to elide registers allocated by intermediate stack frames. |
| 241 | m_globalVarStorageOffset = -RegisterFile::CallFrameHeaderSize - m_codeBlock->m_numParameters - registerFile->size(); |
| 242 | |
| 243 | // Add previously defined symbols to bookkeeping. |
| 244 | m_globals.grow(size: symbolTable->size()); |
| 245 | SymbolTable::iterator end = symbolTable->end(); |
| 246 | for (SymbolTable::iterator it = symbolTable->begin(); it != end; ++it) |
| 247 | registerFor(index: it->second.getIndex()).setIndex(it->second.getIndex() + m_globalVarStorageOffset); |
| 248 | |
| 249 | BatchedTransitionOptimizer optimizer(globalObject); |
| 250 | |
| 251 | const VarStack& varStack = programNode->varStack(); |
| 252 | const FunctionStack& functionStack = programNode->functionStack(); |
| 253 | bool canOptimizeNewGlobals = symbolTable->size() + functionStack.size() + varStack.size() < registerFile->maxGlobals(); |
| 254 | if (canOptimizeNewGlobals) { |
| 255 | // Shift new symbols so they get stored prior to existing symbols. |
| 256 | m_nextGlobalIndex -= symbolTable->size(); |
| 257 | |
| 258 | for (size_t i = 0; i < functionStack.size(); ++i) { |
| 259 | FunctionBodyNode* function = functionStack[i]; |
| 260 | globalObject->removeDirect(propertyName: function->ident()); // Make sure our new function is not shadowed by an old property. |
| 261 | emitNewFunction(dst: addGlobalVar(ident: function->ident(), isConstant: false), body: function); |
| 262 | } |
| 263 | |
| 264 | Vector<RegisterID*, 32> newVars; |
| 265 | for (size_t i = 0; i < varStack.size(); ++i) |
| 266 | if (!globalObject->hasProperty(exec, propertyName: *varStack[i].first)) |
| 267 | newVars.append(val: addGlobalVar(ident: *varStack[i].first, isConstant: varStack[i].second & DeclarationStacks::IsConstant)); |
| 268 | |
| 269 | preserveLastVar(); |
| 270 | |
| 271 | for (size_t i = 0; i < newVars.size(); ++i) |
| 272 | emitLoad(dst: newVars[i], jsUndefined()); |
| 273 | } else { |
| 274 | for (size_t i = 0; i < functionStack.size(); ++i) { |
| 275 | FunctionBodyNode* function = functionStack[i]; |
| 276 | globalObject->putWithAttributes(exec, propertyName: function->ident(), value: new (exec) JSFunction(exec, makeFunction(exec, body: function), scopeChain.node()), attributes: DontDelete); |
| 277 | } |
| 278 | for (size_t i = 0; i < varStack.size(); ++i) { |
| 279 | if (globalObject->hasProperty(exec, propertyName: *varStack[i].first)) |
| 280 | continue; |
| 281 | int attributes = DontDelete; |
| 282 | if (varStack[i].second & DeclarationStacks::IsConstant) |
| 283 | attributes |= ReadOnly; |
| 284 | globalObject->putWithAttributes(exec, propertyName: *varStack[i].first, value: jsUndefined(), attributes); |
| 285 | } |
| 286 | |
| 287 | preserveLastVar(); |
| 288 | } |
| 289 | } |
| 290 | |
| 291 | BytecodeGenerator::BytecodeGenerator(FunctionBodyNode* functionBody, const Debugger* debugger, const ScopeChain& scopeChain, SymbolTable* symbolTable, CodeBlock* codeBlock) |
| 292 | : m_shouldEmitDebugHooks(!!debugger) |
| 293 | , m_shouldEmitProfileHooks(scopeChain.globalObject()->supportsProfiling()) |
| 294 | , m_scopeChain(&scopeChain) |
| 295 | , m_symbolTable(symbolTable) |
| 296 | , m_scopeNode(functionBody) |
| 297 | , m_codeBlock(codeBlock) |
| 298 | , m_finallyDepth(0) |
| 299 | , m_dynamicScopeDepth(0) |
| 300 | , m_baseScopeDepth(0) |
| 301 | , m_codeType(FunctionCode) |
| 302 | , m_nextConstantOffset(0) |
| 303 | , m_globalConstantIndex(0) |
| 304 | , m_globalData(&scopeChain.globalObject()->globalExec()->globalData()) |
| 305 | , m_lastOpcodeID(op_end) |
| 306 | , m_emitNodeDepth(0) |
| 307 | , m_regeneratingForExceptionInfo(false) |
| 308 | , m_codeBlockBeingRegeneratedFrom(0) |
| 309 | { |
| 310 | if (m_shouldEmitDebugHooks) |
| 311 | m_codeBlock->setNeedsFullScopeChain(true); |
| 312 | |
| 313 | codeBlock->setGlobalData(m_globalData); |
| 314 | |
| 315 | bool usesArguments = functionBody->usesArguments(); |
| 316 | codeBlock->setUsesArguments(usesArguments); |
| 317 | if (usesArguments) { |
| 318 | m_argumentsRegister.setIndex(RegisterFile::OptionalCalleeArguments); |
| 319 | addVar(ident: propertyNames().arguments, isConstant: false); |
| 320 | } |
| 321 | |
| 322 | if (m_codeBlock->needsFullScopeChain()) { |
| 323 | ++m_codeBlock->m_numVars; |
| 324 | m_activationRegisterIndex = newRegister()->index(); |
| 325 | emitOpcode(op_enter_with_activation); |
| 326 | instructions().append(val: m_activationRegisterIndex); |
| 327 | } else |
| 328 | emitOpcode(op_enter); |
| 329 | |
| 330 | if (usesArguments) { |
| 331 | emitOpcode(op_init_arguments); |
| 332 | |
| 333 | // The debugger currently retrieves the arguments object from an activation rather than pulling |
| 334 | // it from a call frame. In the long-term it should stop doing that (<rdar://problem/6911886>), |
| 335 | // but for now we force eager creation of the arguments object when debugging. |
| 336 | if (m_shouldEmitDebugHooks) |
| 337 | emitOpcode(op_create_arguments); |
| 338 | } |
| 339 | |
| 340 | const DeclarationStacks::FunctionStack& functionStack = functionBody->functionStack(); |
| 341 | for (size_t i = 0; i < functionStack.size(); ++i) { |
| 342 | FunctionBodyNode* function = functionStack[i]; |
| 343 | const Identifier& ident = function->ident(); |
| 344 | m_functions.add(value: ident.ustring().rep()); |
| 345 | emitNewFunction(dst: addVar(ident, isConstant: false), body: function); |
| 346 | } |
| 347 | |
| 348 | const DeclarationStacks::VarStack& varStack = functionBody->varStack(); |
| 349 | for (size_t i = 0; i < varStack.size(); ++i) |
| 350 | addVar(ident: *varStack[i].first, isConstant: varStack[i].second & DeclarationStacks::IsConstant); |
| 351 | |
| 352 | FunctionParameters& parameters = *functionBody->parameters(); |
| 353 | size_t parameterCount = parameters.size(); |
| 354 | m_nextParameterIndex = -RegisterFile::CallFrameHeaderSize - parameterCount - 1; |
| 355 | m_parameters.grow(size: 1 + parameterCount); // reserve space for "this" |
| 356 | |
| 357 | // Add "this" as a parameter |
| 358 | m_thisRegister.setIndex(m_nextParameterIndex); |
| 359 | ++m_nextParameterIndex; |
| 360 | ++m_codeBlock->m_numParameters; |
| 361 | |
| 362 | if (functionBody->usesThis() || m_shouldEmitDebugHooks) { |
| 363 | emitOpcode(op_convert_this); |
| 364 | instructions().append(val: m_thisRegister.index()); |
| 365 | } |
| 366 | |
| 367 | for (size_t i = 0; i < parameterCount; ++i) |
| 368 | addParameter(parameters[i]); |
| 369 | |
| 370 | preserveLastVar(); |
| 371 | } |
| 372 | |
| 373 | BytecodeGenerator::BytecodeGenerator(EvalNode* evalNode, const Debugger* debugger, const ScopeChain& scopeChain, SymbolTable* symbolTable, EvalCodeBlock* codeBlock) |
| 374 | : m_shouldEmitDebugHooks(!!debugger) |
| 375 | , m_shouldEmitProfileHooks(scopeChain.globalObject()->supportsProfiling()) |
| 376 | , m_scopeChain(&scopeChain) |
| 377 | , m_symbolTable(symbolTable) |
| 378 | , m_scopeNode(evalNode) |
| 379 | , m_codeBlock(codeBlock) |
| 380 | , m_thisRegister(RegisterFile::ProgramCodeThisRegister) |
| 381 | , m_finallyDepth(0) |
| 382 | , m_dynamicScopeDepth(0) |
| 383 | , m_baseScopeDepth(codeBlock->baseScopeDepth()) |
| 384 | , m_codeType(EvalCode) |
| 385 | , m_nextConstantOffset(0) |
| 386 | , m_globalConstantIndex(0) |
| 387 | , m_globalData(&scopeChain.globalObject()->globalExec()->globalData()) |
| 388 | , m_lastOpcodeID(op_end) |
| 389 | , m_emitNodeDepth(0) |
| 390 | , m_regeneratingForExceptionInfo(false) |
| 391 | , m_codeBlockBeingRegeneratedFrom(0) |
| 392 | { |
| 393 | if (m_shouldEmitDebugHooks || m_baseScopeDepth) |
| 394 | m_codeBlock->setNeedsFullScopeChain(true); |
| 395 | |
| 396 | emitOpcode(op_enter); |
| 397 | codeBlock->setGlobalData(m_globalData); |
| 398 | m_codeBlock->m_numParameters = 1; // Allocate space for "this" |
| 399 | |
| 400 | const DeclarationStacks::FunctionStack& functionStack = evalNode->functionStack(); |
| 401 | for (size_t i = 0; i < functionStack.size(); ++i) |
| 402 | m_codeBlock->addFunctionDecl(n: makeFunction(globalData: m_globalData, body: functionStack[i])); |
| 403 | |
| 404 | const DeclarationStacks::VarStack& varStack = evalNode->varStack(); |
| 405 | unsigned numVariables = varStack.size(); |
| 406 | Vector<Identifier> variables; |
| 407 | variables.reserveCapacity(newCapacity: numVariables); |
| 408 | for (size_t i = 0; i < numVariables; ++i) |
| 409 | variables.append(val: *varStack[i].first); |
| 410 | codeBlock->adoptVariables(variables); |
| 411 | |
| 412 | preserveLastVar(); |
| 413 | } |
| 414 | |
| 415 | RegisterID* BytecodeGenerator::addParameter(const Identifier& ident) |
| 416 | { |
| 417 | // Parameters overwrite var declarations, but not function declarations. |
| 418 | RegisterID* result = 0; |
| 419 | UString::Rep* rep = ident.ustring().rep(); |
| 420 | if (!m_functions.contains(value: rep)) { |
| 421 | symbolTable().set(key: rep, mapped: m_nextParameterIndex); |
| 422 | RegisterID& parameter = registerFor(index: m_nextParameterIndex); |
| 423 | parameter.setIndex(m_nextParameterIndex); |
| 424 | result = ¶meter; |
| 425 | } |
| 426 | |
| 427 | // To maintain the calling convention, we have to allocate unique space for |
| 428 | // each parameter, even if the parameter doesn't make it into the symbol table. |
| 429 | ++m_nextParameterIndex; |
| 430 | ++m_codeBlock->m_numParameters; |
| 431 | return result; |
| 432 | } |
| 433 | |
| 434 | RegisterID* BytecodeGenerator::registerFor(const Identifier& ident) |
| 435 | { |
| 436 | if (ident == propertyNames().thisIdentifier) |
| 437 | return &m_thisRegister; |
| 438 | |
| 439 | if (!shouldOptimizeLocals()) |
| 440 | return 0; |
| 441 | |
| 442 | SymbolTableEntry entry = symbolTable().get(key: ident.ustring().rep()); |
| 443 | if (entry.isNull()) |
| 444 | return 0; |
| 445 | |
| 446 | if (ident == propertyNames().arguments) |
| 447 | createArgumentsIfNecessary(); |
| 448 | |
| 449 | return ®isterFor(index: entry.getIndex()); |
| 450 | } |
| 451 | |
| 452 | bool BytecodeGenerator::willResolveToArguments(const Identifier& ident) |
| 453 | { |
| 454 | if (ident != propertyNames().arguments) |
| 455 | return false; |
| 456 | |
| 457 | if (!shouldOptimizeLocals()) |
| 458 | return false; |
| 459 | |
| 460 | SymbolTableEntry entry = symbolTable().get(key: ident.ustring().rep()); |
| 461 | if (entry.isNull()) |
| 462 | return false; |
| 463 | |
| 464 | if (m_codeBlock->usesArguments() && m_codeType == FunctionCode) |
| 465 | return true; |
| 466 | |
| 467 | return false; |
| 468 | } |
| 469 | |
| 470 | RegisterID* BytecodeGenerator::uncheckedRegisterForArguments() |
| 471 | { |
| 472 | ASSERT(willResolveToArguments(propertyNames().arguments)); |
| 473 | |
| 474 | SymbolTableEntry entry = symbolTable().get(key: propertyNames().arguments.ustring().rep()); |
| 475 | ASSERT(!entry.isNull()); |
| 476 | return ®isterFor(index: entry.getIndex()); |
| 477 | } |
| 478 | |
| 479 | RegisterID* BytecodeGenerator::constRegisterFor(const Identifier& ident) |
| 480 | { |
| 481 | if (m_codeType == EvalCode) |
| 482 | return 0; |
| 483 | |
| 484 | SymbolTableEntry entry = symbolTable().get(key: ident.ustring().rep()); |
| 485 | if (entry.isNull()) |
| 486 | return 0; |
| 487 | |
| 488 | return ®isterFor(index: entry.getIndex()); |
| 489 | } |
| 490 | |
| 491 | bool BytecodeGenerator::isLocal(const Identifier& ident) |
| 492 | { |
| 493 | if (ident == propertyNames().thisIdentifier) |
| 494 | return true; |
| 495 | |
| 496 | return shouldOptimizeLocals() && symbolTable().contains(key: ident.ustring().rep()); |
| 497 | } |
| 498 | |
| 499 | bool BytecodeGenerator::isLocalConstant(const Identifier& ident) |
| 500 | { |
| 501 | return symbolTable().get(key: ident.ustring().rep()).isReadOnly(); |
| 502 | } |
| 503 | |
| 504 | RegisterID* BytecodeGenerator::newRegister() |
| 505 | { |
| 506 | m_calleeRegisters.append(value: m_calleeRegisters.size()); |
| 507 | m_codeBlock->m_numCalleeRegisters = max<int>(a: m_codeBlock->m_numCalleeRegisters, b: m_calleeRegisters.size()); |
| 508 | return &m_calleeRegisters.last(); |
| 509 | } |
| 510 | |
| 511 | RegisterID* BytecodeGenerator::newTemporary() |
| 512 | { |
| 513 | // Reclaim free register IDs. |
| 514 | while (m_calleeRegisters.size() && !m_calleeRegisters.last().refCount()) |
| 515 | m_calleeRegisters.removeLast(); |
| 516 | |
| 517 | RegisterID* result = newRegister(); |
| 518 | result->setTemporary(); |
| 519 | return result; |
| 520 | } |
| 521 | |
| 522 | RegisterID* BytecodeGenerator::highestUsedRegister() |
| 523 | { |
| 524 | size_t count = m_codeBlock->m_numCalleeRegisters; |
| 525 | while (m_calleeRegisters.size() < count) |
| 526 | newRegister(); |
| 527 | return &m_calleeRegisters.last(); |
| 528 | } |
| 529 | |
| 530 | PassRefPtr<LabelScope> BytecodeGenerator::newLabelScope(LabelScope::Type type, const Identifier* name) |
| 531 | { |
| 532 | // Reclaim free label scopes. |
| 533 | while (m_labelScopes.size() && !m_labelScopes.last().refCount()) |
| 534 | m_labelScopes.removeLast(); |
| 535 | |
| 536 | // Allocate new label scope. |
| 537 | LabelScope scope(type, name, scopeDepth(), newLabel(), type == LabelScope::Loop ? newLabel() : PassRefPtr<Label>()); // Only loops have continue targets. |
| 538 | m_labelScopes.append(value: scope); |
| 539 | return &m_labelScopes.last(); |
| 540 | } |
| 541 | |
| 542 | PassRefPtr<Label> BytecodeGenerator::newLabel() |
| 543 | { |
| 544 | // Reclaim free label IDs. |
| 545 | while (m_labels.size() && !m_labels.last().refCount()) |
| 546 | m_labels.removeLast(); |
| 547 | |
| 548 | // Allocate new label ID. |
| 549 | m_labels.append(value: m_codeBlock); |
| 550 | return &m_labels.last(); |
| 551 | } |
| 552 | |
| 553 | PassRefPtr<Label> BytecodeGenerator::emitLabel(Label* l0) |
| 554 | { |
| 555 | unsigned newLabelIndex = instructions().size(); |
| 556 | l0->setLocation(newLabelIndex); |
| 557 | |
| 558 | if (m_codeBlock->numberOfJumpTargets()) { |
| 559 | unsigned lastLabelIndex = m_codeBlock->lastJumpTarget(); |
| 560 | ASSERT(lastLabelIndex <= newLabelIndex); |
| 561 | if (newLabelIndex == lastLabelIndex) { |
| 562 | // Peephole optimizations have already been disabled by emitting the last label |
| 563 | return l0; |
| 564 | } |
| 565 | } |
| 566 | |
| 567 | m_codeBlock->addJumpTarget(jumpTarget: newLabelIndex); |
| 568 | |
| 569 | // This disables peephole optimizations when an instruction is a jump target |
| 570 | m_lastOpcodeID = op_end; |
| 571 | return l0; |
| 572 | } |
| 573 | |
| 574 | void BytecodeGenerator::emitOpcode(OpcodeID opcodeID) |
| 575 | { |
| 576 | instructions().append(val: globalData()->interpreter->getOpcode(id: opcodeID)); |
| 577 | m_lastOpcodeID = opcodeID; |
| 578 | } |
| 579 | |
| 580 | void BytecodeGenerator::retrieveLastBinaryOp(int& dstIndex, int& src1Index, int& src2Index) |
| 581 | { |
| 582 | ASSERT(instructions().size() >= 4); |
| 583 | size_t size = instructions().size(); |
| 584 | dstIndex = instructions().at(i: size - 3).u.operand; |
| 585 | src1Index = instructions().at(i: size - 2).u.operand; |
| 586 | src2Index = instructions().at(i: size - 1).u.operand; |
| 587 | } |
| 588 | |
| 589 | void BytecodeGenerator::retrieveLastUnaryOp(int& dstIndex, int& srcIndex) |
| 590 | { |
| 591 | ASSERT(instructions().size() >= 3); |
| 592 | size_t size = instructions().size(); |
| 593 | dstIndex = instructions().at(i: size - 2).u.operand; |
| 594 | srcIndex = instructions().at(i: size - 1).u.operand; |
| 595 | } |
| 596 | |
| 597 | void ALWAYS_INLINE BytecodeGenerator::rewindBinaryOp() |
| 598 | { |
| 599 | ASSERT(instructions().size() >= 4); |
| 600 | instructions().shrink(size: instructions().size() - 4); |
| 601 | } |
| 602 | |
| 603 | void ALWAYS_INLINE BytecodeGenerator::rewindUnaryOp() |
| 604 | { |
| 605 | ASSERT(instructions().size() >= 3); |
| 606 | instructions().shrink(size: instructions().size() - 3); |
| 607 | } |
| 608 | |
| 609 | PassRefPtr<Label> BytecodeGenerator::emitJump(Label* target) |
| 610 | { |
| 611 | size_t begin = instructions().size(); |
| 612 | emitOpcode(opcodeID: target->isForward() ? op_jmp : op_loop); |
| 613 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 614 | return target; |
| 615 | } |
| 616 | |
| 617 | PassRefPtr<Label> BytecodeGenerator::emitJumpIfTrue(RegisterID* cond, Label* target) |
| 618 | { |
| 619 | if (m_lastOpcodeID == op_less) { |
| 620 | int dstIndex; |
| 621 | int src1Index; |
| 622 | int src2Index; |
| 623 | |
| 624 | retrieveLastBinaryOp(dstIndex, src1Index, src2Index); |
| 625 | |
| 626 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 627 | rewindBinaryOp(); |
| 628 | |
| 629 | size_t begin = instructions().size(); |
| 630 | emitOpcode(opcodeID: target->isForward() ? op_jless : op_loop_if_less); |
| 631 | instructions().append(val: src1Index); |
| 632 | instructions().append(val: src2Index); |
| 633 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 634 | return target; |
| 635 | } |
| 636 | } else if (m_lastOpcodeID == op_lesseq && !target->isForward()) { |
| 637 | int dstIndex; |
| 638 | int src1Index; |
| 639 | int src2Index; |
| 640 | |
| 641 | retrieveLastBinaryOp(dstIndex, src1Index, src2Index); |
| 642 | |
| 643 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 644 | rewindBinaryOp(); |
| 645 | |
| 646 | size_t begin = instructions().size(); |
| 647 | emitOpcode(opcodeID: op_loop_if_lesseq); |
| 648 | instructions().append(val: src1Index); |
| 649 | instructions().append(val: src2Index); |
| 650 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 651 | return target; |
| 652 | } |
| 653 | } else if (m_lastOpcodeID == op_eq_null && target->isForward()) { |
| 654 | int dstIndex; |
| 655 | int srcIndex; |
| 656 | |
| 657 | retrieveLastUnaryOp(dstIndex, srcIndex); |
| 658 | |
| 659 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 660 | rewindUnaryOp(); |
| 661 | |
| 662 | size_t begin = instructions().size(); |
| 663 | emitOpcode(opcodeID: op_jeq_null); |
| 664 | instructions().append(val: srcIndex); |
| 665 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 666 | return target; |
| 667 | } |
| 668 | } else if (m_lastOpcodeID == op_neq_null && target->isForward()) { |
| 669 | int dstIndex; |
| 670 | int srcIndex; |
| 671 | |
| 672 | retrieveLastUnaryOp(dstIndex, srcIndex); |
| 673 | |
| 674 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 675 | rewindUnaryOp(); |
| 676 | |
| 677 | size_t begin = instructions().size(); |
| 678 | emitOpcode(opcodeID: op_jneq_null); |
| 679 | instructions().append(val: srcIndex); |
| 680 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 681 | return target; |
| 682 | } |
| 683 | } |
| 684 | |
| 685 | size_t begin = instructions().size(); |
| 686 | |
| 687 | emitOpcode(opcodeID: target->isForward() ? op_jtrue : op_loop_if_true); |
| 688 | instructions().append(val: cond->index()); |
| 689 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 690 | return target; |
| 691 | } |
| 692 | |
| 693 | PassRefPtr<Label> BytecodeGenerator::emitJumpIfFalse(RegisterID* cond, Label* target) |
| 694 | { |
| 695 | if (m_lastOpcodeID == op_less && target->isForward()) { |
| 696 | int dstIndex; |
| 697 | int src1Index; |
| 698 | int src2Index; |
| 699 | |
| 700 | retrieveLastBinaryOp(dstIndex, src1Index, src2Index); |
| 701 | |
| 702 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 703 | rewindBinaryOp(); |
| 704 | |
| 705 | size_t begin = instructions().size(); |
| 706 | emitOpcode(opcodeID: op_jnless); |
| 707 | instructions().append(val: src1Index); |
| 708 | instructions().append(val: src2Index); |
| 709 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 710 | return target; |
| 711 | } |
| 712 | } else if (m_lastOpcodeID == op_lesseq && target->isForward()) { |
| 713 | int dstIndex; |
| 714 | int src1Index; |
| 715 | int src2Index; |
| 716 | |
| 717 | retrieveLastBinaryOp(dstIndex, src1Index, src2Index); |
| 718 | |
| 719 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 720 | rewindBinaryOp(); |
| 721 | |
| 722 | size_t begin = instructions().size(); |
| 723 | emitOpcode(opcodeID: op_jnlesseq); |
| 724 | instructions().append(val: src1Index); |
| 725 | instructions().append(val: src2Index); |
| 726 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 727 | return target; |
| 728 | } |
| 729 | } else if (m_lastOpcodeID == op_not) { |
| 730 | int dstIndex; |
| 731 | int srcIndex; |
| 732 | |
| 733 | retrieveLastUnaryOp(dstIndex, srcIndex); |
| 734 | |
| 735 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 736 | rewindUnaryOp(); |
| 737 | |
| 738 | size_t begin = instructions().size(); |
| 739 | emitOpcode(opcodeID: target->isForward() ? op_jtrue : op_loop_if_true); |
| 740 | instructions().append(val: srcIndex); |
| 741 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 742 | return target; |
| 743 | } |
| 744 | } else if (m_lastOpcodeID == op_eq_null && target->isForward()) { |
| 745 | int dstIndex; |
| 746 | int srcIndex; |
| 747 | |
| 748 | retrieveLastUnaryOp(dstIndex, srcIndex); |
| 749 | |
| 750 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 751 | rewindUnaryOp(); |
| 752 | |
| 753 | size_t begin = instructions().size(); |
| 754 | emitOpcode(opcodeID: op_jneq_null); |
| 755 | instructions().append(val: srcIndex); |
| 756 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 757 | return target; |
| 758 | } |
| 759 | } else if (m_lastOpcodeID == op_neq_null && target->isForward()) { |
| 760 | int dstIndex; |
| 761 | int srcIndex; |
| 762 | |
| 763 | retrieveLastUnaryOp(dstIndex, srcIndex); |
| 764 | |
| 765 | if (cond->index() == dstIndex && cond->isTemporary() && !cond->refCount()) { |
| 766 | rewindUnaryOp(); |
| 767 | |
| 768 | size_t begin = instructions().size(); |
| 769 | emitOpcode(opcodeID: op_jeq_null); |
| 770 | instructions().append(val: srcIndex); |
| 771 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 772 | return target; |
| 773 | } |
| 774 | } |
| 775 | |
| 776 | size_t begin = instructions().size(); |
| 777 | emitOpcode(opcodeID: target->isForward() ? op_jfalse : op_loop_if_false); |
| 778 | instructions().append(val: cond->index()); |
| 779 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 780 | return target; |
| 781 | } |
| 782 | |
| 783 | PassRefPtr<Label> BytecodeGenerator::emitJumpIfNotFunctionCall(RegisterID* cond, Label* target) |
| 784 | { |
| 785 | size_t begin = instructions().size(); |
| 786 | |
| 787 | emitOpcode(opcodeID: op_jneq_ptr); |
| 788 | instructions().append(val: cond->index()); |
| 789 | instructions().append(val: m_scopeChain->globalObject()->d()->callFunction); |
| 790 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 791 | return target; |
| 792 | } |
| 793 | |
| 794 | PassRefPtr<Label> BytecodeGenerator::emitJumpIfNotFunctionApply(RegisterID* cond, Label* target) |
| 795 | { |
| 796 | size_t begin = instructions().size(); |
| 797 | |
| 798 | emitOpcode(opcodeID: op_jneq_ptr); |
| 799 | instructions().append(val: cond->index()); |
| 800 | instructions().append(val: m_scopeChain->globalObject()->d()->applyFunction); |
| 801 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 802 | return target; |
| 803 | } |
| 804 | |
| 805 | unsigned BytecodeGenerator::addConstant(const Identifier& ident) |
| 806 | { |
| 807 | UString::Rep* rep = ident.ustring().rep(); |
| 808 | pair<IdentifierMap::iterator, bool> result = m_identifierMap.add(key: rep, mapped: m_codeBlock->numberOfIdentifiers()); |
| 809 | if (result.second) // new entry |
| 810 | m_codeBlock->addIdentifier(i: Identifier(m_globalData, rep)); |
| 811 | |
| 812 | return result.first->second; |
| 813 | } |
| 814 | |
| 815 | RegisterID* BytecodeGenerator::addConstantValue(JSValue v) |
| 816 | { |
| 817 | int index = m_nextConstantOffset; |
| 818 | |
| 819 | pair<JSValueMap::iterator, bool> result = m_jsValueMap.add(key: JSValue::encode(value: v), mapped: m_nextConstantOffset); |
| 820 | if (result.second) { |
| 821 | m_constantPoolRegisters.append(value: FirstConstantRegisterIndex + m_nextConstantOffset); |
| 822 | ++m_nextConstantOffset; |
| 823 | m_codeBlock->addConstantRegister(r: JSValue(v)); |
| 824 | } else |
| 825 | index = result.first->second; |
| 826 | |
| 827 | return &m_constantPoolRegisters[index]; |
| 828 | } |
| 829 | |
| 830 | unsigned BytecodeGenerator::addRegExp(RegExp* r) |
| 831 | { |
| 832 | return m_codeBlock->addRegExp(r); |
| 833 | } |
| 834 | |
| 835 | RegisterID* BytecodeGenerator::emitMove(RegisterID* dst, RegisterID* src) |
| 836 | { |
| 837 | emitOpcode(opcodeID: op_mov); |
| 838 | instructions().append(val: dst->index()); |
| 839 | instructions().append(val: src->index()); |
| 840 | return dst; |
| 841 | } |
| 842 | |
| 843 | RegisterID* BytecodeGenerator::emitUnaryOp(OpcodeID opcodeID, RegisterID* dst, RegisterID* src) |
| 844 | { |
| 845 | emitOpcode(opcodeID); |
| 846 | instructions().append(val: dst->index()); |
| 847 | instructions().append(val: src->index()); |
| 848 | return dst; |
| 849 | } |
| 850 | |
| 851 | RegisterID* BytecodeGenerator::emitPreInc(RegisterID* srcDst) |
| 852 | { |
| 853 | emitOpcode(opcodeID: op_pre_inc); |
| 854 | instructions().append(val: srcDst->index()); |
| 855 | return srcDst; |
| 856 | } |
| 857 | |
| 858 | RegisterID* BytecodeGenerator::emitPreDec(RegisterID* srcDst) |
| 859 | { |
| 860 | emitOpcode(opcodeID: op_pre_dec); |
| 861 | instructions().append(val: srcDst->index()); |
| 862 | return srcDst; |
| 863 | } |
| 864 | |
| 865 | RegisterID* BytecodeGenerator::emitPostInc(RegisterID* dst, RegisterID* srcDst) |
| 866 | { |
| 867 | emitOpcode(opcodeID: op_post_inc); |
| 868 | instructions().append(val: dst->index()); |
| 869 | instructions().append(val: srcDst->index()); |
| 870 | return dst; |
| 871 | } |
| 872 | |
| 873 | RegisterID* BytecodeGenerator::emitPostDec(RegisterID* dst, RegisterID* srcDst) |
| 874 | { |
| 875 | emitOpcode(opcodeID: op_post_dec); |
| 876 | instructions().append(val: dst->index()); |
| 877 | instructions().append(val: srcDst->index()); |
| 878 | return dst; |
| 879 | } |
| 880 | |
| 881 | RegisterID* BytecodeGenerator::emitBinaryOp(OpcodeID opcodeID, RegisterID* dst, RegisterID* src1, RegisterID* src2, OperandTypes types) |
| 882 | { |
| 883 | emitOpcode(opcodeID); |
| 884 | instructions().append(val: dst->index()); |
| 885 | instructions().append(val: src1->index()); |
| 886 | instructions().append(val: src2->index()); |
| 887 | |
| 888 | if (opcodeID == op_bitor || opcodeID == op_bitand || opcodeID == op_bitxor || |
| 889 | opcodeID == op_add || opcodeID == op_mul || opcodeID == op_sub || opcodeID == op_div) |
| 890 | instructions().append(val: types.toInt()); |
| 891 | |
| 892 | return dst; |
| 893 | } |
| 894 | |
| 895 | RegisterID* BytecodeGenerator::emitEqualityOp(OpcodeID opcodeID, RegisterID* dst, RegisterID* src1, RegisterID* src2) |
| 896 | { |
| 897 | if (m_lastOpcodeID == op_typeof) { |
| 898 | int dstIndex; |
| 899 | int srcIndex; |
| 900 | |
| 901 | retrieveLastUnaryOp(dstIndex, srcIndex); |
| 902 | |
| 903 | if (src1->index() == dstIndex |
| 904 | && src1->isTemporary() |
| 905 | && m_codeBlock->isConstantRegisterIndex(index: src2->index()) |
| 906 | && m_codeBlock->constantRegister(index: src2->index()).jsValue().isString()) { |
| 907 | const UString& value = asString(value: m_codeBlock->constantRegister(index: src2->index()).jsValue())->tryGetValue(); |
| 908 | if (value == "undefined" ) { |
| 909 | rewindUnaryOp(); |
| 910 | emitOpcode(opcodeID: op_is_undefined); |
| 911 | instructions().append(val: dst->index()); |
| 912 | instructions().append(val: srcIndex); |
| 913 | return dst; |
| 914 | } |
| 915 | if (value == "boolean" ) { |
| 916 | rewindUnaryOp(); |
| 917 | emitOpcode(opcodeID: op_is_boolean); |
| 918 | instructions().append(val: dst->index()); |
| 919 | instructions().append(val: srcIndex); |
| 920 | return dst; |
| 921 | } |
| 922 | if (value == "number" ) { |
| 923 | rewindUnaryOp(); |
| 924 | emitOpcode(opcodeID: op_is_number); |
| 925 | instructions().append(val: dst->index()); |
| 926 | instructions().append(val: srcIndex); |
| 927 | return dst; |
| 928 | } |
| 929 | if (value == "string" ) { |
| 930 | rewindUnaryOp(); |
| 931 | emitOpcode(opcodeID: op_is_string); |
| 932 | instructions().append(val: dst->index()); |
| 933 | instructions().append(val: srcIndex); |
| 934 | return dst; |
| 935 | } |
| 936 | if (value == "object" ) { |
| 937 | rewindUnaryOp(); |
| 938 | emitOpcode(opcodeID: op_is_object); |
| 939 | instructions().append(val: dst->index()); |
| 940 | instructions().append(val: srcIndex); |
| 941 | return dst; |
| 942 | } |
| 943 | if (value == "function" ) { |
| 944 | rewindUnaryOp(); |
| 945 | emitOpcode(opcodeID: op_is_function); |
| 946 | instructions().append(val: dst->index()); |
| 947 | instructions().append(val: srcIndex); |
| 948 | return dst; |
| 949 | } |
| 950 | } |
| 951 | } |
| 952 | |
| 953 | emitOpcode(opcodeID); |
| 954 | instructions().append(val: dst->index()); |
| 955 | instructions().append(val: src1->index()); |
| 956 | instructions().append(val: src2->index()); |
| 957 | return dst; |
| 958 | } |
| 959 | |
| 960 | RegisterID* BytecodeGenerator::emitLoad(RegisterID* dst, bool b) |
| 961 | { |
| 962 | return emitLoad(dst, jsBoolean(b)); |
| 963 | } |
| 964 | |
| 965 | RegisterID* BytecodeGenerator::emitLoad(RegisterID* dst, double number) |
| 966 | { |
| 967 | // FIXME: Our hash tables won't hold infinity, so we make a new JSNumberCell each time. |
| 968 | // Later we can do the extra work to handle that like the other cases. |
| 969 | if (number == HashTraits<double>::emptyValue() || HashTraits<double>::isDeletedValue(value: number)) |
| 970 | return emitLoad(dst, jsNumber(globalData: globalData(), d: number)); |
| 971 | JSValue& valueInMap = m_numberMap.add(key: number, mapped: JSValue()).first->second; |
| 972 | if (!valueInMap) |
| 973 | valueInMap = jsNumber(globalData: globalData(), d: number); |
| 974 | return emitLoad(dst, valueInMap); |
| 975 | } |
| 976 | |
| 977 | RegisterID* BytecodeGenerator::emitLoad(RegisterID* dst, const Identifier& identifier) |
| 978 | { |
| 979 | JSString*& stringInMap = m_stringMap.add(key: identifier.ustring().rep(), mapped: 0).first->second; |
| 980 | if (!stringInMap) |
| 981 | stringInMap = jsOwnedString(globalData: globalData(), s: identifier.ustring()); |
| 982 | return emitLoad(dst, JSValue(stringInMap)); |
| 983 | } |
| 984 | |
| 985 | RegisterID* BytecodeGenerator::emitLoad(RegisterID* dst, JSValue v) |
| 986 | { |
| 987 | RegisterID* constantID = addConstantValue(v); |
| 988 | if (dst) |
| 989 | return emitMove(dst, src: constantID); |
| 990 | return constantID; |
| 991 | } |
| 992 | |
| 993 | bool BytecodeGenerator::findScopedProperty(const Identifier& property, int& index, size_t& stackDepth, bool forWriting, JSObject*& globalObject) |
| 994 | { |
| 995 | // Cases where we cannot statically optimize the lookup. |
| 996 | if (property == propertyNames().arguments || !canOptimizeNonLocals()) { |
| 997 | stackDepth = 0; |
| 998 | index = missingSymbolMarker(); |
| 999 | |
| 1000 | if (shouldOptimizeLocals() && m_codeType == GlobalCode) { |
| 1001 | ScopeChainIterator iter = m_scopeChain->begin(); |
| 1002 | globalObject = *iter; |
| 1003 | ASSERT((++iter) == m_scopeChain->end()); |
| 1004 | } |
| 1005 | return false; |
| 1006 | } |
| 1007 | |
| 1008 | size_t depth = 0; |
| 1009 | |
| 1010 | ScopeChainIterator iter = m_scopeChain->begin(); |
| 1011 | ScopeChainIterator end = m_scopeChain->end(); |
| 1012 | for (; iter != end; ++iter, ++depth) { |
| 1013 | JSObject* currentScope = *iter; |
| 1014 | if (!currentScope->isVariableObject()) |
| 1015 | break; |
| 1016 | JSVariableObject* currentVariableObject = static_cast<JSVariableObject*>(currentScope); |
| 1017 | SymbolTableEntry entry = currentVariableObject->symbolTable().get(key: property.ustring().rep()); |
| 1018 | |
| 1019 | // Found the property |
| 1020 | if (!entry.isNull()) { |
| 1021 | if (entry.isReadOnly() && forWriting) { |
| 1022 | stackDepth = 0; |
| 1023 | index = missingSymbolMarker(); |
| 1024 | if (++iter == end) |
| 1025 | globalObject = currentVariableObject; |
| 1026 | return false; |
| 1027 | } |
| 1028 | stackDepth = depth; |
| 1029 | index = entry.getIndex(); |
| 1030 | if (++iter == end) |
| 1031 | globalObject = currentVariableObject; |
| 1032 | return true; |
| 1033 | } |
| 1034 | if (currentVariableObject->isDynamicScope()) |
| 1035 | break; |
| 1036 | } |
| 1037 | |
| 1038 | // Can't locate the property but we're able to avoid a few lookups. |
| 1039 | stackDepth = depth; |
| 1040 | index = missingSymbolMarker(); |
| 1041 | JSObject* scope = *iter; |
| 1042 | if (++iter == end) |
| 1043 | globalObject = scope; |
| 1044 | return true; |
| 1045 | } |
| 1046 | |
| 1047 | RegisterID* BytecodeGenerator::emitInstanceOf(RegisterID* dst, RegisterID* value, RegisterID* base, RegisterID* basePrototype) |
| 1048 | { |
| 1049 | emitOpcode(opcodeID: op_instanceof); |
| 1050 | instructions().append(val: dst->index()); |
| 1051 | instructions().append(val: value->index()); |
| 1052 | instructions().append(val: base->index()); |
| 1053 | instructions().append(val: basePrototype->index()); |
| 1054 | return dst; |
| 1055 | } |
| 1056 | |
| 1057 | RegisterID* BytecodeGenerator::emitResolve(RegisterID* dst, const Identifier& property) |
| 1058 | { |
| 1059 | size_t depth = 0; |
| 1060 | int index = 0; |
| 1061 | JSObject* globalObject = 0; |
| 1062 | if (!findScopedProperty(property, index, stackDepth&: depth, forWriting: false, globalObject) && !globalObject) { |
| 1063 | // We can't optimise at all :-( |
| 1064 | emitOpcode(opcodeID: op_resolve); |
| 1065 | instructions().append(val: dst->index()); |
| 1066 | instructions().append(val: addConstant(ident: property)); |
| 1067 | return dst; |
| 1068 | } |
| 1069 | |
| 1070 | if (globalObject) { |
| 1071 | bool forceGlobalResolve = false; |
| 1072 | if (m_regeneratingForExceptionInfo) { |
| 1073 | #if ENABLE(JIT) |
| 1074 | forceGlobalResolve = m_codeBlockBeingRegeneratedFrom->hasGlobalResolveInfoAtBytecodeOffset(bytecodeOffset: instructions().size()); |
| 1075 | #else |
| 1076 | forceGlobalResolve = m_codeBlockBeingRegeneratedFrom->hasGlobalResolveInstructionAtBytecodeOffset(instructions().size()); |
| 1077 | #endif |
| 1078 | } |
| 1079 | |
| 1080 | if (index != missingSymbolMarker() && !forceGlobalResolve) { |
| 1081 | // Directly index the property lookup across multiple scopes. |
| 1082 | return emitGetScopedVar(dst, skip: depth, index, globalObject); |
| 1083 | } |
| 1084 | |
| 1085 | #if ENABLE(JIT) |
| 1086 | m_codeBlock->addGlobalResolveInfo(globalResolveInstruction: instructions().size()); |
| 1087 | #else |
| 1088 | m_codeBlock->addGlobalResolveInstruction(instructions().size()); |
| 1089 | #endif |
| 1090 | emitOpcode(opcodeID: op_resolve_global); |
| 1091 | instructions().append(val: dst->index()); |
| 1092 | instructions().append(val: globalObject); |
| 1093 | instructions().append(val: addConstant(ident: property)); |
| 1094 | instructions().append(val: 0); |
| 1095 | instructions().append(val: 0); |
| 1096 | return dst; |
| 1097 | } |
| 1098 | |
| 1099 | if (index != missingSymbolMarker()) { |
| 1100 | // Directly index the property lookup across multiple scopes. |
| 1101 | return emitGetScopedVar(dst, skip: depth, index, globalObject); |
| 1102 | } |
| 1103 | |
| 1104 | // In this case we are at least able to drop a few scope chains from the |
| 1105 | // lookup chain, although we still need to hash from then on. |
| 1106 | emitOpcode(opcodeID: op_resolve_skip); |
| 1107 | instructions().append(val: dst->index()); |
| 1108 | instructions().append(val: addConstant(ident: property)); |
| 1109 | instructions().append(val: depth); |
| 1110 | return dst; |
| 1111 | } |
| 1112 | |
| 1113 | RegisterID* BytecodeGenerator::emitGetScopedVar(RegisterID* dst, size_t depth, int index, JSValue globalObject) |
| 1114 | { |
| 1115 | if (globalObject) { |
| 1116 | emitOpcode(opcodeID: op_get_global_var); |
| 1117 | instructions().append(val: dst->index()); |
| 1118 | instructions().append(val: asCell(value: globalObject)); |
| 1119 | instructions().append(val: index); |
| 1120 | return dst; |
| 1121 | } |
| 1122 | |
| 1123 | emitOpcode(opcodeID: op_get_scoped_var); |
| 1124 | instructions().append(val: dst->index()); |
| 1125 | instructions().append(val: index); |
| 1126 | instructions().append(val: depth); |
| 1127 | return dst; |
| 1128 | } |
| 1129 | |
| 1130 | RegisterID* BytecodeGenerator::emitPutScopedVar(size_t depth, int index, RegisterID* value, JSValue globalObject) |
| 1131 | { |
| 1132 | if (globalObject) { |
| 1133 | emitOpcode(opcodeID: op_put_global_var); |
| 1134 | instructions().append(val: asCell(value: globalObject)); |
| 1135 | instructions().append(val: index); |
| 1136 | instructions().append(val: value->index()); |
| 1137 | return value; |
| 1138 | } |
| 1139 | emitOpcode(opcodeID: op_put_scoped_var); |
| 1140 | instructions().append(val: index); |
| 1141 | instructions().append(val: depth); |
| 1142 | instructions().append(val: value->index()); |
| 1143 | return value; |
| 1144 | } |
| 1145 | |
| 1146 | RegisterID* BytecodeGenerator::emitResolveBase(RegisterID* dst, const Identifier& property) |
| 1147 | { |
| 1148 | size_t depth = 0; |
| 1149 | int index = 0; |
| 1150 | JSObject* globalObject = 0; |
| 1151 | findScopedProperty(property, index, stackDepth&: depth, forWriting: false, globalObject); |
| 1152 | if (!globalObject) { |
| 1153 | // We can't optimise at all :-( |
| 1154 | emitOpcode(opcodeID: op_resolve_base); |
| 1155 | instructions().append(val: dst->index()); |
| 1156 | instructions().append(val: addConstant(ident: property)); |
| 1157 | return dst; |
| 1158 | } |
| 1159 | |
| 1160 | // Global object is the base |
| 1161 | return emitLoad(dst, v: JSValue(globalObject)); |
| 1162 | } |
| 1163 | |
| 1164 | RegisterID* BytecodeGenerator::emitResolveWithBase(RegisterID* baseDst, RegisterID* propDst, const Identifier& property) |
| 1165 | { |
| 1166 | size_t depth = 0; |
| 1167 | int index = 0; |
| 1168 | JSObject* globalObject = 0; |
| 1169 | if (!findScopedProperty(property, index, stackDepth&: depth, forWriting: false, globalObject) || !globalObject) { |
| 1170 | // We can't optimise at all :-( |
| 1171 | emitOpcode(opcodeID: op_resolve_with_base); |
| 1172 | instructions().append(val: baseDst->index()); |
| 1173 | instructions().append(val: propDst->index()); |
| 1174 | instructions().append(val: addConstant(ident: property)); |
| 1175 | return baseDst; |
| 1176 | } |
| 1177 | |
| 1178 | bool forceGlobalResolve = false; |
| 1179 | if (m_regeneratingForExceptionInfo) { |
| 1180 | #if ENABLE(JIT) |
| 1181 | forceGlobalResolve = m_codeBlockBeingRegeneratedFrom->hasGlobalResolveInfoAtBytecodeOffset(bytecodeOffset: instructions().size()); |
| 1182 | #else |
| 1183 | forceGlobalResolve = m_codeBlockBeingRegeneratedFrom->hasGlobalResolveInstructionAtBytecodeOffset(instructions().size()); |
| 1184 | #endif |
| 1185 | } |
| 1186 | |
| 1187 | // Global object is the base |
| 1188 | emitLoad(dst: baseDst, v: JSValue(globalObject)); |
| 1189 | |
| 1190 | if (index != missingSymbolMarker() && !forceGlobalResolve) { |
| 1191 | // Directly index the property lookup across multiple scopes. |
| 1192 | emitGetScopedVar(dst: propDst, depth, index, globalObject); |
| 1193 | return baseDst; |
| 1194 | } |
| 1195 | |
| 1196 | #if ENABLE(JIT) |
| 1197 | m_codeBlock->addGlobalResolveInfo(globalResolveInstruction: instructions().size()); |
| 1198 | #else |
| 1199 | m_codeBlock->addGlobalResolveInstruction(instructions().size()); |
| 1200 | #endif |
| 1201 | emitOpcode(opcodeID: op_resolve_global); |
| 1202 | instructions().append(val: propDst->index()); |
| 1203 | instructions().append(val: globalObject); |
| 1204 | instructions().append(val: addConstant(ident: property)); |
| 1205 | instructions().append(val: 0); |
| 1206 | instructions().append(val: 0); |
| 1207 | return baseDst; |
| 1208 | } |
| 1209 | |
| 1210 | void BytecodeGenerator::emitMethodCheck() |
| 1211 | { |
| 1212 | emitOpcode(opcodeID: op_method_check); |
| 1213 | } |
| 1214 | |
| 1215 | RegisterID* BytecodeGenerator::emitGetById(RegisterID* dst, RegisterID* base, const Identifier& property) |
| 1216 | { |
| 1217 | #if ENABLE(JIT) |
| 1218 | m_codeBlock->addStructureStubInfo(stubInfo: StructureStubInfo(access_get_by_id)); |
| 1219 | #else |
| 1220 | m_codeBlock->addPropertyAccessInstruction(instructions().size()); |
| 1221 | #endif |
| 1222 | |
| 1223 | emitOpcode(opcodeID: op_get_by_id); |
| 1224 | instructions().append(val: dst->index()); |
| 1225 | instructions().append(val: base->index()); |
| 1226 | instructions().append(val: addConstant(ident: property)); |
| 1227 | instructions().append(val: 0); |
| 1228 | instructions().append(val: 0); |
| 1229 | instructions().append(val: 0); |
| 1230 | instructions().append(val: 0); |
| 1231 | return dst; |
| 1232 | } |
| 1233 | |
| 1234 | RegisterID* BytecodeGenerator::emitPutById(RegisterID* base, const Identifier& property, RegisterID* value) |
| 1235 | { |
| 1236 | #if ENABLE(JIT) |
| 1237 | m_codeBlock->addStructureStubInfo(stubInfo: StructureStubInfo(access_put_by_id)); |
| 1238 | #else |
| 1239 | m_codeBlock->addPropertyAccessInstruction(instructions().size()); |
| 1240 | #endif |
| 1241 | |
| 1242 | emitOpcode(opcodeID: op_put_by_id); |
| 1243 | instructions().append(val: base->index()); |
| 1244 | instructions().append(val: addConstant(ident: property)); |
| 1245 | instructions().append(val: value->index()); |
| 1246 | instructions().append(val: 0); |
| 1247 | instructions().append(val: 0); |
| 1248 | instructions().append(val: 0); |
| 1249 | instructions().append(val: 0); |
| 1250 | return value; |
| 1251 | } |
| 1252 | |
| 1253 | RegisterID* BytecodeGenerator::emitPutGetter(RegisterID* base, const Identifier& property, RegisterID* value) |
| 1254 | { |
| 1255 | emitOpcode(opcodeID: op_put_getter); |
| 1256 | instructions().append(val: base->index()); |
| 1257 | instructions().append(val: addConstant(ident: property)); |
| 1258 | instructions().append(val: value->index()); |
| 1259 | return value; |
| 1260 | } |
| 1261 | |
| 1262 | RegisterID* BytecodeGenerator::emitPutSetter(RegisterID* base, const Identifier& property, RegisterID* value) |
| 1263 | { |
| 1264 | emitOpcode(opcodeID: op_put_setter); |
| 1265 | instructions().append(val: base->index()); |
| 1266 | instructions().append(val: addConstant(ident: property)); |
| 1267 | instructions().append(val: value->index()); |
| 1268 | return value; |
| 1269 | } |
| 1270 | |
| 1271 | RegisterID* BytecodeGenerator::emitDeleteById(RegisterID* dst, RegisterID* base, const Identifier& property) |
| 1272 | { |
| 1273 | emitOpcode(opcodeID: op_del_by_id); |
| 1274 | instructions().append(val: dst->index()); |
| 1275 | instructions().append(val: base->index()); |
| 1276 | instructions().append(val: addConstant(ident: property)); |
| 1277 | return dst; |
| 1278 | } |
| 1279 | |
| 1280 | RegisterID* BytecodeGenerator::emitGetByVal(RegisterID* dst, RegisterID* base, RegisterID* property) |
| 1281 | { |
| 1282 | for (size_t i = m_forInContextStack.size(); i > 0; i--) { |
| 1283 | ForInContext& context = m_forInContextStack[i - 1]; |
| 1284 | if (context.propertyRegister == property) { |
| 1285 | emitOpcode(opcodeID: op_get_by_pname); |
| 1286 | instructions().append(val: dst->index()); |
| 1287 | instructions().append(val: base->index()); |
| 1288 | instructions().append(val: property->index()); |
| 1289 | instructions().append(val: context.expectedSubscriptRegister->index()); |
| 1290 | instructions().append(val: context.iterRegister->index()); |
| 1291 | instructions().append(val: context.indexRegister->index()); |
| 1292 | return dst; |
| 1293 | } |
| 1294 | } |
| 1295 | emitOpcode(opcodeID: op_get_by_val); |
| 1296 | instructions().append(val: dst->index()); |
| 1297 | instructions().append(val: base->index()); |
| 1298 | instructions().append(val: property->index()); |
| 1299 | return dst; |
| 1300 | } |
| 1301 | |
| 1302 | RegisterID* BytecodeGenerator::emitPutByVal(RegisterID* base, RegisterID* property, RegisterID* value) |
| 1303 | { |
| 1304 | emitOpcode(opcodeID: op_put_by_val); |
| 1305 | instructions().append(val: base->index()); |
| 1306 | instructions().append(val: property->index()); |
| 1307 | instructions().append(val: value->index()); |
| 1308 | return value; |
| 1309 | } |
| 1310 | |
| 1311 | RegisterID* BytecodeGenerator::emitDeleteByVal(RegisterID* dst, RegisterID* base, RegisterID* property) |
| 1312 | { |
| 1313 | emitOpcode(opcodeID: op_del_by_val); |
| 1314 | instructions().append(val: dst->index()); |
| 1315 | instructions().append(val: base->index()); |
| 1316 | instructions().append(val: property->index()); |
| 1317 | return dst; |
| 1318 | } |
| 1319 | |
| 1320 | RegisterID* BytecodeGenerator::emitPutByIndex(RegisterID* base, unsigned index, RegisterID* value) |
| 1321 | { |
| 1322 | emitOpcode(opcodeID: op_put_by_index); |
| 1323 | instructions().append(val: base->index()); |
| 1324 | instructions().append(val: index); |
| 1325 | instructions().append(val: value->index()); |
| 1326 | return value; |
| 1327 | } |
| 1328 | |
| 1329 | RegisterID* BytecodeGenerator::emitNewObject(RegisterID* dst) |
| 1330 | { |
| 1331 | emitOpcode(opcodeID: op_new_object); |
| 1332 | instructions().append(val: dst->index()); |
| 1333 | return dst; |
| 1334 | } |
| 1335 | |
| 1336 | RegisterID* BytecodeGenerator::emitNewArray(RegisterID* dst, ElementNode* elements) |
| 1337 | { |
| 1338 | Vector<RefPtr<RegisterID>, 16> argv; |
| 1339 | for (ElementNode* n = elements; n; n = n->next()) { |
| 1340 | if (n->elision()) |
| 1341 | break; |
| 1342 | argv.append(val: newTemporary()); |
| 1343 | // op_new_array requires the initial values to be a sequential range of registers |
| 1344 | ASSERT(argv.size() == 1 || argv[argv.size() - 1]->index() == argv[argv.size() - 2]->index() + 1); |
| 1345 | emitNode(dst: argv.last().get(), n: n->value()); |
| 1346 | } |
| 1347 | emitOpcode(opcodeID: op_new_array); |
| 1348 | instructions().append(val: dst->index()); |
| 1349 | instructions().append(val: argv.size() ? argv[0]->index() : 0); // argv |
| 1350 | instructions().append(val: argv.size()); // argc |
| 1351 | return dst; |
| 1352 | } |
| 1353 | |
| 1354 | RegisterID* BytecodeGenerator::emitNewFunction(RegisterID* dst, FunctionBodyNode* function) |
| 1355 | { |
| 1356 | unsigned index = m_codeBlock->addFunctionDecl(n: makeFunction(globalData: m_globalData, body: function)); |
| 1357 | |
| 1358 | emitOpcode(opcodeID: op_new_func); |
| 1359 | instructions().append(val: dst->index()); |
| 1360 | instructions().append(val: index); |
| 1361 | return dst; |
| 1362 | } |
| 1363 | |
| 1364 | RegisterID* BytecodeGenerator::emitNewRegExp(RegisterID* dst, RegExp* regExp) |
| 1365 | { |
| 1366 | emitOpcode(opcodeID: op_new_regexp); |
| 1367 | instructions().append(val: dst->index()); |
| 1368 | instructions().append(val: addRegExp(r: regExp)); |
| 1369 | return dst; |
| 1370 | } |
| 1371 | |
| 1372 | |
| 1373 | RegisterID* BytecodeGenerator::emitNewFunctionExpression(RegisterID* r0, FuncExprNode* n) |
| 1374 | { |
| 1375 | FunctionBodyNode* function = n->body(); |
| 1376 | unsigned index = m_codeBlock->addFunctionExpr(n: makeFunction(globalData: m_globalData, body: function)); |
| 1377 | |
| 1378 | emitOpcode(opcodeID: op_new_func_exp); |
| 1379 | instructions().append(val: r0->index()); |
| 1380 | instructions().append(val: index); |
| 1381 | return r0; |
| 1382 | } |
| 1383 | |
| 1384 | RegisterID* BytecodeGenerator::emitCall(RegisterID* dst, RegisterID* func, RegisterID* thisRegister, ArgumentsNode* argumentsNode, unsigned divot, unsigned startOffset, unsigned endOffset) |
| 1385 | { |
| 1386 | return emitCall(op_call, dst, func, thisRegister, argumentsNode, divot, startOffset, endOffset); |
| 1387 | } |
| 1388 | |
| 1389 | void BytecodeGenerator::createArgumentsIfNecessary() |
| 1390 | { |
| 1391 | if (m_codeBlock->usesArguments() && m_codeType == FunctionCode) |
| 1392 | emitOpcode(opcodeID: op_create_arguments); |
| 1393 | } |
| 1394 | |
| 1395 | RegisterID* BytecodeGenerator::emitCallEval(RegisterID* dst, RegisterID* func, RegisterID* thisRegister, ArgumentsNode* argumentsNode, unsigned divot, unsigned startOffset, unsigned endOffset) |
| 1396 | { |
| 1397 | createArgumentsIfNecessary(); |
| 1398 | return emitCall(op_call_eval, dst, func, thisRegister, argumentsNode, divot, startOffset, endOffset); |
| 1399 | } |
| 1400 | |
| 1401 | RegisterID* BytecodeGenerator::emitCall(OpcodeID opcodeID, RegisterID* dst, RegisterID* func, RegisterID* thisRegister, ArgumentsNode* argumentsNode, unsigned divot, unsigned startOffset, unsigned endOffset) |
| 1402 | { |
| 1403 | ASSERT(opcodeID == op_call || opcodeID == op_call_eval); |
| 1404 | ASSERT(func->refCount()); |
| 1405 | ASSERT(thisRegister->refCount()); |
| 1406 | |
| 1407 | RegisterID* originalFunc = func; |
| 1408 | if (m_shouldEmitProfileHooks) { |
| 1409 | // If codegen decided to recycle func as this call's destination register, |
| 1410 | // we need to undo that optimization here so that func will still be around |
| 1411 | // for the sake of op_profile_did_call. |
| 1412 | if (dst == func) { |
| 1413 | RefPtr<RegisterID> movedThisRegister = emitMove(dst: newTemporary(), src: thisRegister); |
| 1414 | RefPtr<RegisterID> movedFunc = emitMove(dst: thisRegister, src: func); |
| 1415 | |
| 1416 | thisRegister = movedThisRegister.release().releaseRef(); |
| 1417 | func = movedFunc.release().releaseRef(); |
| 1418 | } |
| 1419 | } |
| 1420 | |
| 1421 | // Generate code for arguments. |
| 1422 | Vector<RefPtr<RegisterID>, 16> argv; |
| 1423 | argv.append(val: thisRegister); |
| 1424 | for (ArgumentListNode* n = argumentsNode->m_listNode; n; n = n->m_next) { |
| 1425 | argv.append(val: newTemporary()); |
| 1426 | // op_call requires the arguments to be a sequential range of registers |
| 1427 | ASSERT(argv[argv.size() - 1]->index() == argv[argv.size() - 2]->index() + 1); |
| 1428 | emitNode(dst: argv.last().get(), n); |
| 1429 | } |
| 1430 | |
| 1431 | // Reserve space for call frame. |
| 1432 | Vector<RefPtr<RegisterID>, RegisterFile::CallFrameHeaderSize> callFrame; |
| 1433 | for (int i = 0; i < RegisterFile::CallFrameHeaderSize; ++i) |
| 1434 | callFrame.append(val: newTemporary()); |
| 1435 | |
| 1436 | if (m_shouldEmitProfileHooks) { |
| 1437 | emitOpcode(opcodeID: op_profile_will_call); |
| 1438 | instructions().append(val: func->index()); |
| 1439 | |
| 1440 | #if ENABLE(JIT) |
| 1441 | m_codeBlock->addFunctionRegisterInfo(bytecodeOffset: instructions().size(), functionIndex: func->index()); |
| 1442 | #endif |
| 1443 | } |
| 1444 | |
| 1445 | emitExpressionInfo(divot, startOffset, endOffset); |
| 1446 | |
| 1447 | #if ENABLE(JIT) |
| 1448 | m_codeBlock->addCallLinkInfo(); |
| 1449 | #endif |
| 1450 | |
| 1451 | // Emit call. |
| 1452 | emitOpcode(opcodeID); |
| 1453 | instructions().append(val: dst->index()); // dst |
| 1454 | instructions().append(val: func->index()); // func |
| 1455 | instructions().append(val: argv.size()); // argCount |
| 1456 | instructions().append(val: argv[0]->index() + argv.size() + RegisterFile::CallFrameHeaderSize); // registerOffset |
| 1457 | |
| 1458 | if (m_shouldEmitProfileHooks) { |
| 1459 | emitOpcode(opcodeID: op_profile_did_call); |
| 1460 | instructions().append(val: func->index()); |
| 1461 | |
| 1462 | if (dst == originalFunc) { |
| 1463 | thisRegister->deref(); |
| 1464 | func->deref(); |
| 1465 | } |
| 1466 | } |
| 1467 | |
| 1468 | return dst; |
| 1469 | } |
| 1470 | |
| 1471 | RegisterID* BytecodeGenerator::emitLoadVarargs(RegisterID* argCountDst, RegisterID* arguments) |
| 1472 | { |
| 1473 | ASSERT(argCountDst->index() < arguments->index()); |
| 1474 | emitOpcode(opcodeID: op_load_varargs); |
| 1475 | instructions().append(val: argCountDst->index()); |
| 1476 | instructions().append(val: arguments->index()); |
| 1477 | return argCountDst; |
| 1478 | } |
| 1479 | |
| 1480 | RegisterID* BytecodeGenerator::emitCallVarargs(RegisterID* dst, RegisterID* func, RegisterID* thisRegister, RegisterID* argCountRegister, unsigned divot, unsigned startOffset, unsigned endOffset) |
| 1481 | { |
| 1482 | ASSERT(func->refCount()); |
| 1483 | ASSERT(thisRegister->refCount()); |
| 1484 | ASSERT(dst != func); |
| 1485 | if (m_shouldEmitProfileHooks) { |
| 1486 | emitOpcode(opcodeID: op_profile_will_call); |
| 1487 | instructions().append(val: func->index()); |
| 1488 | |
| 1489 | #if ENABLE(JIT) |
| 1490 | m_codeBlock->addFunctionRegisterInfo(bytecodeOffset: instructions().size(), functionIndex: func->index()); |
| 1491 | #endif |
| 1492 | } |
| 1493 | |
| 1494 | emitExpressionInfo(divot, startOffset, endOffset); |
| 1495 | |
| 1496 | // Emit call. |
| 1497 | emitOpcode(opcodeID: op_call_varargs); |
| 1498 | instructions().append(val: dst->index()); // dst |
| 1499 | instructions().append(val: func->index()); // func |
| 1500 | instructions().append(val: argCountRegister->index()); // arg count |
| 1501 | instructions().append(val: thisRegister->index() + RegisterFile::CallFrameHeaderSize); // initial registerOffset |
| 1502 | if (m_shouldEmitProfileHooks) { |
| 1503 | emitOpcode(opcodeID: op_profile_did_call); |
| 1504 | instructions().append(val: func->index()); |
| 1505 | } |
| 1506 | return dst; |
| 1507 | } |
| 1508 | |
| 1509 | RegisterID* BytecodeGenerator::emitReturn(RegisterID* src) |
| 1510 | { |
| 1511 | if (m_codeBlock->needsFullScopeChain()) { |
| 1512 | emitOpcode(opcodeID: op_tear_off_activation); |
| 1513 | instructions().append(val: m_activationRegisterIndex); |
| 1514 | } else if (m_codeBlock->usesArguments() && m_codeBlock->m_numParameters > 1) |
| 1515 | emitOpcode(opcodeID: op_tear_off_arguments); |
| 1516 | |
| 1517 | return emitUnaryNoDstOp(op_ret, src); |
| 1518 | } |
| 1519 | |
| 1520 | RegisterID* BytecodeGenerator::emitUnaryNoDstOp(OpcodeID opcodeID, RegisterID* src) |
| 1521 | { |
| 1522 | emitOpcode(opcodeID); |
| 1523 | instructions().append(val: src->index()); |
| 1524 | return src; |
| 1525 | } |
| 1526 | |
| 1527 | RegisterID* BytecodeGenerator::emitConstruct(RegisterID* dst, RegisterID* func, ArgumentsNode* argumentsNode, unsigned divot, unsigned startOffset, unsigned endOffset) |
| 1528 | { |
| 1529 | ASSERT(func->refCount()); |
| 1530 | |
| 1531 | RegisterID* originalFunc = func; |
| 1532 | if (m_shouldEmitProfileHooks) { |
| 1533 | // If codegen decided to recycle func as this call's destination register, |
| 1534 | // we need to undo that optimization here so that func will still be around |
| 1535 | // for the sake of op_profile_did_call. |
| 1536 | if (dst == func) { |
| 1537 | RefPtr<RegisterID> movedFunc = emitMove(dst: newTemporary(), src: func); |
| 1538 | func = movedFunc.release().releaseRef(); |
| 1539 | } |
| 1540 | } |
| 1541 | |
| 1542 | RefPtr<RegisterID> funcProto = newTemporary(); |
| 1543 | |
| 1544 | // Generate code for arguments. |
| 1545 | Vector<RefPtr<RegisterID>, 16> argv; |
| 1546 | argv.append(val: newTemporary()); // reserve space for "this" |
| 1547 | for (ArgumentListNode* n = argumentsNode ? argumentsNode->m_listNode : 0; n; n = n->m_next) { |
| 1548 | argv.append(val: newTemporary()); |
| 1549 | // op_construct requires the arguments to be a sequential range of registers |
| 1550 | ASSERT(argv[argv.size() - 1]->index() == argv[argv.size() - 2]->index() + 1); |
| 1551 | emitNode(dst: argv.last().get(), n); |
| 1552 | } |
| 1553 | |
| 1554 | if (m_shouldEmitProfileHooks) { |
| 1555 | emitOpcode(opcodeID: op_profile_will_call); |
| 1556 | instructions().append(val: func->index()); |
| 1557 | } |
| 1558 | |
| 1559 | // Load prototype. |
| 1560 | emitExpressionInfo(divot, startOffset, endOffset); |
| 1561 | emitGetByIdExceptionInfo(opcodeID: op_construct); |
| 1562 | emitGetById(dst: funcProto.get(), base: func, property: globalData()->propertyNames->prototype); |
| 1563 | |
| 1564 | // Reserve space for call frame. |
| 1565 | Vector<RefPtr<RegisterID>, RegisterFile::CallFrameHeaderSize> callFrame; |
| 1566 | for (int i = 0; i < RegisterFile::CallFrameHeaderSize; ++i) |
| 1567 | callFrame.append(val: newTemporary()); |
| 1568 | |
| 1569 | emitExpressionInfo(divot, startOffset, endOffset); |
| 1570 | |
| 1571 | #if ENABLE(JIT) |
| 1572 | m_codeBlock->addCallLinkInfo(); |
| 1573 | #endif |
| 1574 | |
| 1575 | emitOpcode(opcodeID: op_construct); |
| 1576 | instructions().append(val: dst->index()); // dst |
| 1577 | instructions().append(val: func->index()); // func |
| 1578 | instructions().append(val: argv.size()); // argCount |
| 1579 | instructions().append(val: argv[0]->index() + argv.size() + RegisterFile::CallFrameHeaderSize); // registerOffset |
| 1580 | instructions().append(val: funcProto->index()); // proto |
| 1581 | instructions().append(val: argv[0]->index()); // thisRegister |
| 1582 | |
| 1583 | emitOpcode(opcodeID: op_construct_verify); |
| 1584 | instructions().append(val: dst->index()); |
| 1585 | instructions().append(val: argv[0]->index()); |
| 1586 | |
| 1587 | if (m_shouldEmitProfileHooks) { |
| 1588 | emitOpcode(opcodeID: op_profile_did_call); |
| 1589 | instructions().append(val: func->index()); |
| 1590 | |
| 1591 | if (dst == originalFunc) |
| 1592 | func->deref(); |
| 1593 | } |
| 1594 | |
| 1595 | return dst; |
| 1596 | } |
| 1597 | |
| 1598 | RegisterID* BytecodeGenerator::emitStrcat(RegisterID* dst, RegisterID* src, int count) |
| 1599 | { |
| 1600 | emitOpcode(opcodeID: op_strcat); |
| 1601 | instructions().append(val: dst->index()); |
| 1602 | instructions().append(val: src->index()); |
| 1603 | instructions().append(val: count); |
| 1604 | |
| 1605 | return dst; |
| 1606 | } |
| 1607 | |
| 1608 | void BytecodeGenerator::emitToPrimitive(RegisterID* dst, RegisterID* src) |
| 1609 | { |
| 1610 | emitOpcode(opcodeID: op_to_primitive); |
| 1611 | instructions().append(val: dst->index()); |
| 1612 | instructions().append(val: src->index()); |
| 1613 | } |
| 1614 | |
| 1615 | RegisterID* BytecodeGenerator::emitPushScope(RegisterID* scope) |
| 1616 | { |
| 1617 | ASSERT(scope->isTemporary()); |
| 1618 | ControlFlowContext context; |
| 1619 | context.isFinallyBlock = false; |
| 1620 | m_scopeContextStack.append(val: context); |
| 1621 | m_dynamicScopeDepth++; |
| 1622 | createArgumentsIfNecessary(); |
| 1623 | |
| 1624 | return emitUnaryNoDstOp(opcodeID: op_push_scope, src: scope); |
| 1625 | } |
| 1626 | |
| 1627 | void BytecodeGenerator::emitPopScope() |
| 1628 | { |
| 1629 | ASSERT(m_scopeContextStack.size()); |
| 1630 | ASSERT(!m_scopeContextStack.last().isFinallyBlock); |
| 1631 | |
| 1632 | emitOpcode(opcodeID: op_pop_scope); |
| 1633 | |
| 1634 | m_scopeContextStack.removeLast(); |
| 1635 | m_dynamicScopeDepth--; |
| 1636 | } |
| 1637 | |
| 1638 | void BytecodeGenerator::emitDebugHook(DebugHookID debugHookID, int firstLine, int lastLine) |
| 1639 | { |
| 1640 | if (!m_shouldEmitDebugHooks) |
| 1641 | return; |
| 1642 | emitOpcode(opcodeID: op_debug); |
| 1643 | instructions().append(val: debugHookID); |
| 1644 | instructions().append(val: firstLine); |
| 1645 | instructions().append(val: lastLine); |
| 1646 | } |
| 1647 | |
| 1648 | void BytecodeGenerator::pushFinallyContext(Label* target, RegisterID* retAddrDst) |
| 1649 | { |
| 1650 | ControlFlowContext scope; |
| 1651 | scope.isFinallyBlock = true; |
| 1652 | FinallyContext context = { .finallyAddr: target, .retAddrDst: retAddrDst }; |
| 1653 | scope.finallyContext = context; |
| 1654 | m_scopeContextStack.append(val: scope); |
| 1655 | m_finallyDepth++; |
| 1656 | } |
| 1657 | |
| 1658 | void BytecodeGenerator::popFinallyContext() |
| 1659 | { |
| 1660 | ASSERT(m_scopeContextStack.size()); |
| 1661 | ASSERT(m_scopeContextStack.last().isFinallyBlock); |
| 1662 | ASSERT(m_finallyDepth > 0); |
| 1663 | m_scopeContextStack.removeLast(); |
| 1664 | m_finallyDepth--; |
| 1665 | } |
| 1666 | |
| 1667 | LabelScope* BytecodeGenerator::breakTarget(const Identifier& name) |
| 1668 | { |
| 1669 | // Reclaim free label scopes. |
| 1670 | // |
| 1671 | // The condition was previously coded as 'm_labelScopes.size() && !m_labelScopes.last().refCount()', |
| 1672 | // however sometimes this appears to lead to GCC going a little haywire and entering the loop with |
| 1673 | // size 0, leading to segfaulty badness. We are yet to identify a valid cause within our code to |
| 1674 | // cause the GCC codegen to misbehave in this fashion, and as such the following refactoring of the |
| 1675 | // loop condition is a workaround. |
| 1676 | while (m_labelScopes.size()) { |
| 1677 | if (m_labelScopes.last().refCount()) |
| 1678 | break; |
| 1679 | m_labelScopes.removeLast(); |
| 1680 | } |
| 1681 | |
| 1682 | if (!m_labelScopes.size()) |
| 1683 | return 0; |
| 1684 | |
| 1685 | // We special-case the following, which is a syntax error in Firefox: |
| 1686 | // label: |
| 1687 | // break; |
| 1688 | if (name.isEmpty()) { |
| 1689 | for (int i = m_labelScopes.size() - 1; i >= 0; --i) { |
| 1690 | LabelScope* scope = &m_labelScopes[i]; |
| 1691 | if (scope->type() != LabelScope::NamedLabel) { |
| 1692 | ASSERT(scope->breakTarget()); |
| 1693 | return scope; |
| 1694 | } |
| 1695 | } |
| 1696 | return 0; |
| 1697 | } |
| 1698 | |
| 1699 | for (int i = m_labelScopes.size() - 1; i >= 0; --i) { |
| 1700 | LabelScope* scope = &m_labelScopes[i]; |
| 1701 | if (scope->name() && *scope->name() == name) { |
| 1702 | ASSERT(scope->breakTarget()); |
| 1703 | return scope; |
| 1704 | } |
| 1705 | } |
| 1706 | return 0; |
| 1707 | } |
| 1708 | |
| 1709 | LabelScope* BytecodeGenerator::continueTarget(const Identifier& name) |
| 1710 | { |
| 1711 | // Reclaim free label scopes. |
| 1712 | while (m_labelScopes.size() && !m_labelScopes.last().refCount()) |
| 1713 | m_labelScopes.removeLast(); |
| 1714 | |
| 1715 | if (!m_labelScopes.size()) |
| 1716 | return 0; |
| 1717 | |
| 1718 | if (name.isEmpty()) { |
| 1719 | for (int i = m_labelScopes.size() - 1; i >= 0; --i) { |
| 1720 | LabelScope* scope = &m_labelScopes[i]; |
| 1721 | if (scope->type() == LabelScope::Loop) { |
| 1722 | ASSERT(scope->continueTarget()); |
| 1723 | return scope; |
| 1724 | } |
| 1725 | } |
| 1726 | return 0; |
| 1727 | } |
| 1728 | |
| 1729 | // Continue to the loop nested nearest to the label scope that matches |
| 1730 | // 'name'. |
| 1731 | LabelScope* result = 0; |
| 1732 | for (int i = m_labelScopes.size() - 1; i >= 0; --i) { |
| 1733 | LabelScope* scope = &m_labelScopes[i]; |
| 1734 | if (scope->type() == LabelScope::Loop) { |
| 1735 | ASSERT(scope->continueTarget()); |
| 1736 | result = scope; |
| 1737 | } |
| 1738 | if (scope->name() && *scope->name() == name) |
| 1739 | return result; // may be 0 |
| 1740 | } |
| 1741 | return 0; |
| 1742 | } |
| 1743 | |
| 1744 | PassRefPtr<Label> BytecodeGenerator::emitComplexJumpScopes(Label* target, ControlFlowContext* topScope, ControlFlowContext* bottomScope) |
| 1745 | { |
| 1746 | while (topScope > bottomScope) { |
| 1747 | // First we count the number of dynamic scopes we need to remove to get |
| 1748 | // to a finally block. |
| 1749 | int nNormalScopes = 0; |
| 1750 | while (topScope > bottomScope) { |
| 1751 | if (topScope->isFinallyBlock) |
| 1752 | break; |
| 1753 | ++nNormalScopes; |
| 1754 | --topScope; |
| 1755 | } |
| 1756 | |
| 1757 | if (nNormalScopes) { |
| 1758 | size_t begin = instructions().size(); |
| 1759 | |
| 1760 | // We need to remove a number of dynamic scopes to get to the next |
| 1761 | // finally block |
| 1762 | emitOpcode(opcodeID: op_jmp_scopes); |
| 1763 | instructions().append(val: nNormalScopes); |
| 1764 | |
| 1765 | // If topScope == bottomScope then there isn't actually a finally block |
| 1766 | // left to emit, so make the jmp_scopes jump directly to the target label |
| 1767 | if (topScope == bottomScope) { |
| 1768 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 1769 | return target; |
| 1770 | } |
| 1771 | |
| 1772 | // Otherwise we just use jmp_scopes to pop a group of scopes and go |
| 1773 | // to the next instruction |
| 1774 | RefPtr<Label> nextInsn = newLabel(); |
| 1775 | instructions().append(val: nextInsn->bind(opcode: begin, offset: instructions().size())); |
| 1776 | emitLabel(l0: nextInsn.get()); |
| 1777 | } |
| 1778 | |
| 1779 | while (topScope > bottomScope && topScope->isFinallyBlock) { |
| 1780 | emitJumpSubroutine(retAddrDst: topScope->finallyContext.retAddrDst, topScope->finallyContext.finallyAddr); |
| 1781 | --topScope; |
| 1782 | } |
| 1783 | } |
| 1784 | return emitJump(target); |
| 1785 | } |
| 1786 | |
| 1787 | PassRefPtr<Label> BytecodeGenerator::emitJumpScopes(Label* target, int targetScopeDepth) |
| 1788 | { |
| 1789 | ASSERT(scopeDepth() - targetScopeDepth >= 0); |
| 1790 | ASSERT(target->isForward()); |
| 1791 | |
| 1792 | size_t scopeDelta = scopeDepth() - targetScopeDepth; |
| 1793 | ASSERT(scopeDelta <= m_scopeContextStack.size()); |
| 1794 | if (!scopeDelta) |
| 1795 | return emitJump(target); |
| 1796 | |
| 1797 | if (m_finallyDepth) |
| 1798 | return emitComplexJumpScopes(target, topScope: &m_scopeContextStack.last(), bottomScope: &m_scopeContextStack.last() - scopeDelta); |
| 1799 | |
| 1800 | size_t begin = instructions().size(); |
| 1801 | |
| 1802 | emitOpcode(opcodeID: op_jmp_scopes); |
| 1803 | instructions().append(val: scopeDelta); |
| 1804 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 1805 | return target; |
| 1806 | } |
| 1807 | |
| 1808 | RegisterID* BytecodeGenerator::emitGetPropertyNames(RegisterID* dst, RegisterID* base, RegisterID* i, RegisterID* size, Label* breakTarget) |
| 1809 | { |
| 1810 | size_t begin = instructions().size(); |
| 1811 | |
| 1812 | emitOpcode(opcodeID: op_get_pnames); |
| 1813 | instructions().append(val: dst->index()); |
| 1814 | instructions().append(val: base->index()); |
| 1815 | instructions().append(val: i->index()); |
| 1816 | instructions().append(val: size->index()); |
| 1817 | instructions().append(val: breakTarget->bind(opcode: begin, offset: instructions().size())); |
| 1818 | return dst; |
| 1819 | } |
| 1820 | |
| 1821 | RegisterID* BytecodeGenerator::emitNextPropertyName(RegisterID* dst, RegisterID* base, RegisterID* i, RegisterID* size, RegisterID* iter, Label* target) |
| 1822 | { |
| 1823 | size_t begin = instructions().size(); |
| 1824 | |
| 1825 | emitOpcode(opcodeID: op_next_pname); |
| 1826 | instructions().append(val: dst->index()); |
| 1827 | instructions().append(val: base->index()); |
| 1828 | instructions().append(val: i->index()); |
| 1829 | instructions().append(val: size->index()); |
| 1830 | instructions().append(val: iter->index()); |
| 1831 | instructions().append(val: target->bind(opcode: begin, offset: instructions().size())); |
| 1832 | return dst; |
| 1833 | } |
| 1834 | |
| 1835 | RegisterID* BytecodeGenerator::emitCatch(RegisterID* targetRegister, Label* start, Label* end) |
| 1836 | { |
| 1837 | #if ENABLE(JIT) |
| 1838 | HandlerInfo info = { |
| 1839 | .start: static_cast<uint32_t>(start->bind(opcode: 0, offset: 0)), |
| 1840 | .end: static_cast<uint32_t>(end->bind(opcode: 0, offset: 0)), |
| 1841 | .target: static_cast<uint32_t>(instructions().size()), |
| 1842 | .scopeDepth: static_cast<uint32_t>(m_dynamicScopeDepth + m_baseScopeDepth), |
| 1843 | .nativeCode: CodeLocationLabel() |
| 1844 | }; |
| 1845 | #else |
| 1846 | HandlerInfo info = { |
| 1847 | static_cast<uint32_t>(start->bind(0, 0)), |
| 1848 | static_cast<uint32_t>(end->bind(0, 0)), |
| 1849 | static_cast<uint32_t>(instructions().size()), |
| 1850 | static_cast<uint32_t>(m_dynamicScopeDepth + m_baseScopeDepth) |
| 1851 | }; |
| 1852 | #endif |
| 1853 | |
| 1854 | m_codeBlock->addExceptionHandler(hanler: info); |
| 1855 | emitOpcode(opcodeID: op_catch); |
| 1856 | instructions().append(val: targetRegister->index()); |
| 1857 | return targetRegister; |
| 1858 | } |
| 1859 | |
| 1860 | RegisterID* BytecodeGenerator::emitNewError(RegisterID* dst, ErrorType type, JSValue message) |
| 1861 | { |
| 1862 | emitOpcode(opcodeID: op_new_error); |
| 1863 | instructions().append(val: dst->index()); |
| 1864 | instructions().append(val: static_cast<int>(type)); |
| 1865 | instructions().append(val: addConstantValue(v: message)->index()); |
| 1866 | return dst; |
| 1867 | } |
| 1868 | |
| 1869 | PassRefPtr<Label> BytecodeGenerator::emitJumpSubroutine(RegisterID* retAddrDst, Label* finally) |
| 1870 | { |
| 1871 | size_t begin = instructions().size(); |
| 1872 | |
| 1873 | emitOpcode(opcodeID: op_jsr); |
| 1874 | instructions().append(val: retAddrDst->index()); |
| 1875 | instructions().append(val: finally->bind(opcode: begin, offset: instructions().size())); |
| 1876 | emitLabel(l0: newLabel().get()); // Record the fact that the next instruction is implicitly labeled, because op_sret will return to it. |
| 1877 | return finally; |
| 1878 | } |
| 1879 | |
| 1880 | void BytecodeGenerator::emitSubroutineReturn(RegisterID* retAddrSrc) |
| 1881 | { |
| 1882 | emitOpcode(opcodeID: op_sret); |
| 1883 | instructions().append(val: retAddrSrc->index()); |
| 1884 | } |
| 1885 | |
| 1886 | void BytecodeGenerator::emitPushNewScope(RegisterID* dst, const Identifier& property, RegisterID* value) |
| 1887 | { |
| 1888 | ControlFlowContext context; |
| 1889 | context.isFinallyBlock = false; |
| 1890 | m_scopeContextStack.append(val: context); |
| 1891 | m_dynamicScopeDepth++; |
| 1892 | |
| 1893 | createArgumentsIfNecessary(); |
| 1894 | |
| 1895 | emitOpcode(opcodeID: op_push_new_scope); |
| 1896 | instructions().append(val: dst->index()); |
| 1897 | instructions().append(val: addConstant(ident: property)); |
| 1898 | instructions().append(val: value->index()); |
| 1899 | } |
| 1900 | |
| 1901 | void BytecodeGenerator::beginSwitch(RegisterID* scrutineeRegister, SwitchInfo::SwitchType type) |
| 1902 | { |
| 1903 | SwitchInfo info = { .bytecodeOffset: static_cast<uint32_t>(instructions().size()), .switchType: type }; |
| 1904 | switch (type) { |
| 1905 | case SwitchInfo::SwitchImmediate: |
| 1906 | emitOpcode(opcodeID: op_switch_imm); |
| 1907 | break; |
| 1908 | case SwitchInfo::SwitchCharacter: |
| 1909 | emitOpcode(opcodeID: op_switch_char); |
| 1910 | break; |
| 1911 | case SwitchInfo::SwitchString: |
| 1912 | emitOpcode(opcodeID: op_switch_string); |
| 1913 | break; |
| 1914 | default: |
| 1915 | ASSERT_NOT_REACHED(); |
| 1916 | } |
| 1917 | |
| 1918 | instructions().append(val: 0); // place holder for table index |
| 1919 | instructions().append(val: 0); // place holder for default target |
| 1920 | instructions().append(val: scrutineeRegister->index()); |
| 1921 | m_switchContextStack.append(val: info); |
| 1922 | } |
| 1923 | |
| 1924 | static int32_t keyForImmediateSwitch(ExpressionNode* node, int32_t min, int32_t max) |
| 1925 | { |
| 1926 | UNUSED_PARAM(max); |
| 1927 | ASSERT(node->isNumber()); |
| 1928 | double value = static_cast<NumberNode*>(node)->value(); |
| 1929 | int32_t key = static_cast<int32_t>(value); |
| 1930 | ASSERT(key == value); |
| 1931 | ASSERT(key >= min); |
| 1932 | ASSERT(key <= max); |
| 1933 | return key - min; |
| 1934 | } |
| 1935 | |
| 1936 | static void prepareJumpTableForImmediateSwitch(SimpleJumpTable& jumpTable, int32_t switchAddress, uint32_t clauseCount, RefPtr<Label>* labels, ExpressionNode** nodes, int32_t min, int32_t max) |
| 1937 | { |
| 1938 | jumpTable.min = min; |
| 1939 | jumpTable.branchOffsets.resize(size: max - min + 1); |
| 1940 | jumpTable.branchOffsets.fill(val: 0); |
| 1941 | for (uint32_t i = 0; i < clauseCount; ++i) { |
| 1942 | // We're emitting this after the clause labels should have been fixed, so |
| 1943 | // the labels should not be "forward" references |
| 1944 | ASSERT(!labels[i]->isForward()); |
| 1945 | jumpTable.add(key: keyForImmediateSwitch(node: nodes[i], min, max), offset: labels[i]->bind(opcode: switchAddress, offset: switchAddress + 3)); |
| 1946 | } |
| 1947 | } |
| 1948 | |
| 1949 | static int32_t keyForCharacterSwitch(ExpressionNode* node, int32_t min, int32_t max) |
| 1950 | { |
| 1951 | UNUSED_PARAM(max); |
| 1952 | ASSERT(node->isString()); |
| 1953 | UString::Rep* clause = static_cast<StringNode*>(node)->value().ustring().rep(); |
| 1954 | ASSERT(clause->size() == 1); |
| 1955 | |
| 1956 | int32_t key = clause->data()[0]; |
| 1957 | ASSERT(key >= min); |
| 1958 | ASSERT(key <= max); |
| 1959 | return key - min; |
| 1960 | } |
| 1961 | |
| 1962 | static void prepareJumpTableForCharacterSwitch(SimpleJumpTable& jumpTable, int32_t switchAddress, uint32_t clauseCount, RefPtr<Label>* labels, ExpressionNode** nodes, int32_t min, int32_t max) |
| 1963 | { |
| 1964 | jumpTable.min = min; |
| 1965 | jumpTable.branchOffsets.resize(size: max - min + 1); |
| 1966 | jumpTable.branchOffsets.fill(val: 0); |
| 1967 | for (uint32_t i = 0; i < clauseCount; ++i) { |
| 1968 | // We're emitting this after the clause labels should have been fixed, so |
| 1969 | // the labels should not be "forward" references |
| 1970 | ASSERT(!labels[i]->isForward()); |
| 1971 | jumpTable.add(key: keyForCharacterSwitch(node: nodes[i], min, max), offset: labels[i]->bind(opcode: switchAddress, offset: switchAddress + 3)); |
| 1972 | } |
| 1973 | } |
| 1974 | |
| 1975 | static void prepareJumpTableForStringSwitch(StringJumpTable& jumpTable, int32_t switchAddress, uint32_t clauseCount, RefPtr<Label>* labels, ExpressionNode** nodes) |
| 1976 | { |
| 1977 | for (uint32_t i = 0; i < clauseCount; ++i) { |
| 1978 | // We're emitting this after the clause labels should have been fixed, so |
| 1979 | // the labels should not be "forward" references |
| 1980 | ASSERT(!labels[i]->isForward()); |
| 1981 | |
| 1982 | ASSERT(nodes[i]->isString()); |
| 1983 | UString::Rep* clause = static_cast<StringNode*>(nodes[i])->value().ustring().rep(); |
| 1984 | OffsetLocation location; |
| 1985 | location.branchOffset = labels[i]->bind(opcode: switchAddress, offset: switchAddress + 3); |
| 1986 | jumpTable.offsetTable.add(key: clause, mapped: location); |
| 1987 | } |
| 1988 | } |
| 1989 | |
| 1990 | void BytecodeGenerator::endSwitch(uint32_t clauseCount, RefPtr<Label>* labels, ExpressionNode** nodes, Label* defaultLabel, int32_t min, int32_t max) |
| 1991 | { |
| 1992 | SwitchInfo switchInfo = m_switchContextStack.last(); |
| 1993 | m_switchContextStack.removeLast(); |
| 1994 | if (switchInfo.switchType == SwitchInfo::SwitchImmediate) { |
| 1995 | instructions()[switchInfo.bytecodeOffset + 1] = m_codeBlock->numberOfImmediateSwitchJumpTables(); |
| 1996 | instructions()[switchInfo.bytecodeOffset + 2] = defaultLabel->bind(opcode: switchInfo.bytecodeOffset, offset: switchInfo.bytecodeOffset + 3); |
| 1997 | |
| 1998 | SimpleJumpTable& jumpTable = m_codeBlock->addImmediateSwitchJumpTable(); |
| 1999 | prepareJumpTableForImmediateSwitch(jumpTable, switchAddress: switchInfo.bytecodeOffset, clauseCount, labels, nodes, min, max); |
| 2000 | } else if (switchInfo.switchType == SwitchInfo::SwitchCharacter) { |
| 2001 | instructions()[switchInfo.bytecodeOffset + 1] = m_codeBlock->numberOfCharacterSwitchJumpTables(); |
| 2002 | instructions()[switchInfo.bytecodeOffset + 2] = defaultLabel->bind(opcode: switchInfo.bytecodeOffset, offset: switchInfo.bytecodeOffset + 3); |
| 2003 | |
| 2004 | SimpleJumpTable& jumpTable = m_codeBlock->addCharacterSwitchJumpTable(); |
| 2005 | prepareJumpTableForCharacterSwitch(jumpTable, switchAddress: switchInfo.bytecodeOffset, clauseCount, labels, nodes, min, max); |
| 2006 | } else { |
| 2007 | ASSERT(switchInfo.switchType == SwitchInfo::SwitchString); |
| 2008 | instructions()[switchInfo.bytecodeOffset + 1] = m_codeBlock->numberOfStringSwitchJumpTables(); |
| 2009 | instructions()[switchInfo.bytecodeOffset + 2] = defaultLabel->bind(opcode: switchInfo.bytecodeOffset, offset: switchInfo.bytecodeOffset + 3); |
| 2010 | |
| 2011 | StringJumpTable& jumpTable = m_codeBlock->addStringSwitchJumpTable(); |
| 2012 | prepareJumpTableForStringSwitch(jumpTable, switchAddress: switchInfo.bytecodeOffset, clauseCount, labels, nodes); |
| 2013 | } |
| 2014 | } |
| 2015 | |
| 2016 | RegisterID* BytecodeGenerator::emitThrowExpressionTooDeepException() |
| 2017 | { |
| 2018 | // It would be nice to do an even better job of identifying exactly where the expression is. |
| 2019 | // And we could make the caller pass the node pointer in, if there was some way of getting |
| 2020 | // that from an arbitrary node. However, calling emitExpressionInfo without any useful data |
| 2021 | // is still good enough to get us an accurate line number. |
| 2022 | emitExpressionInfo(divot: 0, startOffset: 0, endOffset: 0); |
| 2023 | RegisterID* exception = emitNewError(dst: newTemporary(), type: SyntaxError, message: jsString(globalData: globalData(), s: "Expression too deep" )); |
| 2024 | emitThrow(exc: exception); |
| 2025 | return exception; |
| 2026 | } |
| 2027 | |
| 2028 | } // namespace JSC |
| 2029 | |