Revision control

Copy as Markdown

Other Tools

/*
* Copyright © 2010 Intel Corporation
*
* Permission is hereby granted, free of charge, to any person obtaining a
* copy of this software and associated documentation files (the "Software"),
* to deal in the Software without restriction, including without limitation
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
* and/or sell copies of the Software, and to permit persons to whom the
* Software is furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice (including the next
* paragraph) shall be included in all copies or substantial portions of the
* Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
* DEALINGS IN THE SOFTWARE.
*/
/**
* \file ast_to_hir.c
* Convert abstract syntax to to high-level intermediate reprensentation (HIR).
*
* During the conversion to HIR, the majority of the symantic checking is
* preformed on the program. This includes:
*
* * Symbol table management
* * Type checking
* * Function binding
*
* The majority of this work could be done during parsing, and the parser could
* probably generate HIR directly. However, this results in frequent changes
* to the parser code. Since we do not assume that every system this complier
* is built on will have Flex and Bison installed, we have to store the code
* generated by these tools in our version control system. In other parts of
* the system we've seen problems where a parser was changed but the generated
* code was not committed, merge conflicts where created because two developers
* had slightly different versions of Bison installed, etc.
*
* I have also noticed that running Bison generated parsers in GDB is very
* irritating. When you get a segfault on '$$ = $1->foo', you can't very
* well 'print $1' in GDB.
*
* As a result, my preference is to put as little C code as possible in the
* parser (and lexer) sources.
*/
#include "glsl_symbol_table.h"
#include "glsl_parser_extras.h"
#include "ast.h"
#include "compiler/glsl_types.h"
#include "util/hash_table.h"
#include "main/mtypes.h"
#include "main/macros.h"
#include "main/shaderobj.h"
#include "ir.h"
#include "ir_builder.h"
#include "builtin_functions.h"
using namespace ir_builder;
static void
detect_conflicting_assignments(struct _mesa_glsl_parse_state *state,
exec_list *instructions);
static void
verify_subroutine_associated_funcs(struct _mesa_glsl_parse_state *state);
static void
remove_per_vertex_blocks(exec_list *instructions,
_mesa_glsl_parse_state *state, ir_variable_mode mode);
/**
* Visitor class that finds the first instance of any write-only variable that
* is ever read, if any
*/
class read_from_write_only_variable_visitor : public ir_hierarchical_visitor
{
public:
read_from_write_only_variable_visitor() : found(NULL)
{
}
virtual ir_visitor_status visit(ir_dereference_variable *ir)
{
if (this->in_assignee)
return visit_continue;
ir_variable *var = ir->variable_referenced();
/* We can have memory_write_only set on both images and buffer variables,
* but in the former there is a distinction between reads from
* the variable itself (write_only) and from the memory they point to
* (memory_write_only), while in the case of buffer variables there is
* no such distinction, that is why this check here is limited to
* buffer variables alone.
*/
if (!var || var->data.mode != ir_var_shader_storage)
return visit_continue;
if (var->data.memory_write_only) {
found = var;
return visit_stop;
}
return visit_continue;
}
ir_variable *get_variable() {
return found;
}
virtual ir_visitor_status visit_enter(ir_expression *ir)
{
/* .length() doesn't actually read anything */
if (ir->operation == ir_unop_ssbo_unsized_array_length)
return visit_continue_with_parent;
return visit_continue;
}
private:
ir_variable *found;
};
void
_mesa_ast_to_hir(exec_list *instructions, struct _mesa_glsl_parse_state *state)
{
_mesa_glsl_initialize_variables(instructions, state);
state->symbols->separate_function_namespace = state->language_version == 110;
state->current_function = NULL;
state->toplevel_ir = instructions;
state->gs_input_prim_type_specified = false;
state->tcs_output_vertices_specified = false;
state->cs_input_local_size_specified = false;
/* Section 4.2 of the GLSL 1.20 specification states:
* "The built-in functions are scoped in a scope outside the global scope
* users declare global variables in. That is, a shader's global scope,
* available for user-defined functions and global variables, is nested
* inside the scope containing the built-in functions."
*
* Since built-in functions like ftransform() access built-in variables,
* it follows that those must be in the outer scope as well.
*
* We push scope here to create this nesting effect...but don't pop.
* This way, a shader's globals are still in the symbol table for use
* by the linker.
*/
state->symbols->push_scope();
foreach_list_typed (ast_node, ast, link, & state->translation_unit)
ast->hir(instructions, state);
verify_subroutine_associated_funcs(state);
detect_recursion_unlinked(state, instructions);
detect_conflicting_assignments(state, instructions);
state->toplevel_ir = NULL;
/* Move all of the variable declarations to the front of the IR list, and
* reverse the order. This has the (intended!) side effect that vertex
* shader inputs and fragment shader outputs will appear in the IR in the
* same order that they appeared in the shader code. This results in the
* locations being assigned in the declared order. Many (arguably buggy)
* applications depend on this behavior, and it matches what nearly all
* other drivers do.
* However, do not push the declarations before struct decls or precision
* statements.
*/
ir_instruction* before_node = (ir_instruction*)instructions->get_head();
ir_instruction* after_node = NULL;
while (before_node && (before_node->ir_type == ir_type_precision || before_node->ir_type == ir_type_typedecl))
{
after_node = before_node;
before_node = (ir_instruction*)before_node->next;
}
foreach_in_list_safe(ir_instruction, node, instructions) {
ir_variable *const var = node->as_variable();
if (var == NULL)
continue;
var->remove();
if (after_node)
after_node->insert_after(var);
else
instructions->push_head(var);
}
/* Figure out if gl_FragCoord is actually used in fragment shader */
ir_variable *const var = state->symbols->get_variable("gl_FragCoord");
if (var != NULL)
state->fs_uses_gl_fragcoord = var->data.used;
/* From section 7.1 (Built-In Language Variables) of the GLSL 4.10 spec:
*
* If multiple shaders using members of a built-in block belonging to
* the same interface are linked together in the same program, they
* must all redeclare the built-in block in the same way, as described
* in section 4.3.7 "Interface Blocks" for interface block matching, or
* a link error will result.
*
* The phrase "using members of a built-in block" implies that if two
* shaders are linked together and one of them *does not use* any members
* of the built-in block, then that shader does not need to have a matching
* redeclaration of the built-in block.
*
* This appears to be a clarification to the behaviour established for
* gl_PerVertex by GLSL 1.50, therefore implement it regardless of GLSL
* version.
*
* The definition of "interface" in section 4.3.7 that applies here is as
* follows:
*
* The boundary between adjacent programmable pipeline stages: This
* spans all the outputs in all compilation units of the first stage
* and all the inputs in all compilation units of the second stage.
*
* Therefore this rule applies to both inter- and intra-stage linking.
*
* The easiest way to implement this is to check whether the shader uses
* gl_PerVertex right after ast-to-ir conversion, and if it doesn't, simply
* remove all the relevant variable declaration from the IR, so that the
* linker won't see them and complain about mismatches.
*/
remove_per_vertex_blocks(instructions, state, ir_var_shader_in);
remove_per_vertex_blocks(instructions, state, ir_var_shader_out);
/* Check that we don't have reads from write-only variables */
read_from_write_only_variable_visitor v;
v.run(instructions);
ir_variable *error_var = v.get_variable();
if (error_var) {
/* It would be nice to have proper location information, but for that
* we would need to check this as we process each kind of AST node
*/
YYLTYPE loc;
memset(&loc, 0, sizeof(loc));
_mesa_glsl_error(&loc, state, "Read from write-only variable `%s'",
error_var->name);
}
}
static ir_expression_operation
get_implicit_conversion_operation(const glsl_type *to, const glsl_type *from,
struct _mesa_glsl_parse_state *state)
{
switch (to->base_type) {
case GLSL_TYPE_FLOAT:
switch (from->base_type) {
case GLSL_TYPE_INT: return ir_unop_i2f;
case GLSL_TYPE_UINT: return ir_unop_u2f;
default: return (ir_expression_operation)0;
}
case GLSL_TYPE_UINT:
if (!state->has_implicit_uint_to_int_conversion())
return (ir_expression_operation)0;
switch (from->base_type) {
case GLSL_TYPE_INT: return ir_unop_i2u;
default: return (ir_expression_operation)0;
}
case GLSL_TYPE_DOUBLE:
if (!state->has_double())
return (ir_expression_operation)0;
switch (from->base_type) {
case GLSL_TYPE_INT: return ir_unop_i2d;
case GLSL_TYPE_UINT: return ir_unop_u2d;
case GLSL_TYPE_FLOAT: return ir_unop_f2d;
case GLSL_TYPE_INT64: return ir_unop_i642d;
case GLSL_TYPE_UINT64: return ir_unop_u642d;
default: return (ir_expression_operation)0;
}
case GLSL_TYPE_UINT64:
if (!state->has_int64())
return (ir_expression_operation)0;
switch (from->base_type) {
case GLSL_TYPE_INT: return ir_unop_i2u64;
case GLSL_TYPE_UINT: return ir_unop_u2u64;
case GLSL_TYPE_INT64: return ir_unop_i642u64;
default: return (ir_expression_operation)0;
}
case GLSL_TYPE_INT64:
if (!state->has_int64())
return (ir_expression_operation)0;
switch (from->base_type) {
case GLSL_TYPE_INT: return ir_unop_i2i64;
default: return (ir_expression_operation)0;
}
default: return (ir_expression_operation)0;
}
}
/**
* If a conversion is available, convert one operand to a different type
*
* The \c from \c ir_rvalue is converted "in place".
*
* \param to Type that the operand it to be converted to
* \param from Operand that is being converted
* \param state GLSL compiler state
*
* \return
* If a conversion is possible (or unnecessary), \c true is returned.
* Otherwise \c false is returned.
*/
static bool
apply_implicit_conversion(const glsl_type *to, ir_rvalue * &from,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
if (to->base_type == from->type->base_type)
return true;
/* Prior to GLSL 1.20, there are no implicit conversions */
if (!state->has_implicit_conversions())
return false;
/* From page 27 (page 33 of the PDF) of the GLSL 1.50 spec:
*
* "There are no implicit array or structure conversions. For
* example, an array of int cannot be implicitly converted to an
* array of float.
*/
if (!to->is_numeric() || !from->type->is_numeric())
return false;
/* We don't actually want the specific type `to`, we want a type
* with the same base type as `to`, but the same vector width as
* `from`.
*/
to = glsl_type::get_instance(to->base_type, from->type->vector_elements,
from->type->matrix_columns);
ir_expression_operation op = get_implicit_conversion_operation(to, from->type, state);
if (op) {
from = new(ctx) ir_expression(op, to, from, NULL);
return true;
} else {
return false;
}
}
static const struct glsl_type *
arithmetic_result_type(ir_rvalue * &value_a, ir_rvalue * &value_b,
bool multiply,
struct _mesa_glsl_parse_state *state, YYLTYPE *loc)
{
const glsl_type *type_a = value_a->type;
const glsl_type *type_b = value_b->type;
/* From GLSL 1.50 spec, page 56:
*
* "The arithmetic binary operators add (+), subtract (-),
* multiply (*), and divide (/) operate on integer and
* floating-point scalars, vectors, and matrices."
*/
if (!type_a->is_numeric() || !type_b->is_numeric()) {
_mesa_glsl_error(loc, state,
"operands to arithmetic operators must be numeric");
return glsl_type::error_type;
}
/* "If one operand is floating-point based and the other is
* not, then the conversions from Section 4.1.10 "Implicit
* Conversions" are applied to the non-floating-point-based operand."
*/
if (!apply_implicit_conversion(type_a, value_b, state)
&& !apply_implicit_conversion(type_b, value_a, state)) {
_mesa_glsl_error(loc, state,
"could not implicitly convert operands to "
"arithmetic operator");
return glsl_type::error_type;
}
type_a = value_a->type;
type_b = value_b->type;
/* "If the operands are integer types, they must both be signed or
* both be unsigned."
*
* From this rule and the preceeding conversion it can be inferred that
* both types must be GLSL_TYPE_FLOAT, or GLSL_TYPE_UINT, or GLSL_TYPE_INT.
* The is_numeric check above already filtered out the case where either
* type is not one of these, so now the base types need only be tested for
* equality.
*/
if (type_a->base_type != type_b->base_type) {
_mesa_glsl_error(loc, state,
"base type mismatch for arithmetic operator");
return glsl_type::error_type;
}
/* "All arithmetic binary operators result in the same fundamental type
* (signed integer, unsigned integer, or floating-point) as the
* operands they operate on, after operand type conversion. After
* conversion, the following cases are valid
*
* * The two operands are scalars. In this case the operation is
* applied, resulting in a scalar."
*/
if (type_a->is_scalar() && type_b->is_scalar())
return type_a;
/* "* One operand is a scalar, and the other is a vector or matrix.
* In this case, the scalar operation is applied independently to each
* component of the vector or matrix, resulting in the same size
* vector or matrix."
*/
if (type_a->is_scalar()) {
if (!type_b->is_scalar())
return type_b;
} else if (type_b->is_scalar()) {
return type_a;
}
/* All of the combinations of <scalar, scalar>, <vector, scalar>,
* <scalar, vector>, <scalar, matrix>, and <matrix, scalar> have been
* handled.
*/
assert(!type_a->is_scalar());
assert(!type_b->is_scalar());
/* "* The two operands are vectors of the same size. In this case, the
* operation is done component-wise resulting in the same size
* vector."
*/
if (type_a->is_vector() && type_b->is_vector()) {
if (type_a == type_b) {
return type_a;
} else {
_mesa_glsl_error(loc, state,
"vector size mismatch for arithmetic operator");
return glsl_type::error_type;
}
}
/* All of the combinations of <scalar, scalar>, <vector, scalar>,
* <scalar, vector>, <scalar, matrix>, <matrix, scalar>, and
* <vector, vector> have been handled. At least one of the operands must
* be matrix. Further, since there are no integer matrix types, the base
* type of both operands must be float.
*/
assert(type_a->is_matrix() || type_b->is_matrix());
assert(type_a->is_float() || type_a->is_double());
assert(type_b->is_float() || type_b->is_double());
/* "* The operator is add (+), subtract (-), or divide (/), and the
* operands are matrices with the same number of rows and the same
* number of columns. In this case, the operation is done component-
* wise resulting in the same size matrix."
* * The operator is multiply (*), where both operands are matrices or
* one operand is a vector and the other a matrix. A right vector
* operand is treated as a column vector and a left vector operand as a
* row vector. In all these cases, it is required that the number of
* columns of the left operand is equal to the number of rows of the
* right operand. Then, the multiply (*) operation does a linear
* algebraic multiply, yielding an object that has the same number of
* rows as the left operand and the same number of columns as the right
* operand. Section 5.10 "Vector and Matrix Operations" explains in
* more detail how vectors and matrices are operated on."
*/
if (! multiply) {
if (type_a == type_b)
return type_a;
} else {
const glsl_type *type = glsl_type::get_mul_type(type_a, type_b);
if (type == glsl_type::error_type) {
_mesa_glsl_error(loc, state,
"size mismatch for matrix multiplication");
}
return type;
}
/* "All other cases are illegal."
*/
_mesa_glsl_error(loc, state, "type mismatch");
return glsl_type::error_type;
}
static const struct glsl_type *
unary_arithmetic_result_type(const struct glsl_type *type,
struct _mesa_glsl_parse_state *state, YYLTYPE *loc)
{
/* From GLSL 1.50 spec, page 57:
*
* "The arithmetic unary operators negate (-), post- and pre-increment
* and decrement (-- and ++) operate on integer or floating-point
* values (including vectors and matrices). All unary operators work
* component-wise on their operands. These result with the same type
* they operated on."
*/
if (!type->is_numeric()) {
_mesa_glsl_error(loc, state,
"operands to arithmetic operators must be numeric");
return glsl_type::error_type;
}
return type;
}
/**
* \brief Return the result type of a bit-logic operation.
*
* If the given types to the bit-logic operator are invalid, return
* glsl_type::error_type.
*
* \param value_a LHS of bit-logic op
* \param value_b RHS of bit-logic op
*/
static const struct glsl_type *
bit_logic_result_type(ir_rvalue * &value_a, ir_rvalue * &value_b,
ast_operators op,
struct _mesa_glsl_parse_state *state, YYLTYPE *loc)
{
const glsl_type *type_a = value_a->type;
const glsl_type *type_b = value_b->type;
if (!state->check_bitwise_operations_allowed(loc)) {
return glsl_type::error_type;
}
/* From page 50 (page 56 of PDF) of GLSL 1.30 spec:
*
* "The bitwise operators and (&), exclusive-or (^), and inclusive-or
* (|). The operands must be of type signed or unsigned integers or
* integer vectors."
*/
if (!type_a->is_integer_32_64()) {
_mesa_glsl_error(loc, state, "LHS of `%s' must be an integer",
ast_expression::operator_string(op));
return glsl_type::error_type;
}
if (!type_b->is_integer_32_64()) {
_mesa_glsl_error(loc, state, "RHS of `%s' must be an integer",
ast_expression::operator_string(op));
return glsl_type::error_type;
}
/* Prior to GLSL 4.0 / GL_ARB_gpu_shader5, implicit conversions didn't
* make sense for bitwise operations, as they don't operate on floats.
*
* GLSL 4.0 added implicit int -> uint conversions, which are relevant
* here. It wasn't clear whether or not we should apply them to bitwise
* operations. However, Khronos has decided that they should in future
* language revisions. Applications also rely on this behavior. We opt
* to apply them in general, but issue a portability warning.
*
*/
if (type_a->base_type != type_b->base_type) {
if (!apply_implicit_conversion(type_a, value_b, state)
&& !apply_implicit_conversion(type_b, value_a, state)) {
_mesa_glsl_error(loc, state,
"could not implicitly convert operands to "
"`%s` operator",
ast_expression::operator_string(op));
return glsl_type::error_type;
} else {
_mesa_glsl_warning(loc, state,
"some implementations may not support implicit "
"int -> uint conversions for `%s' operators; "
"consider casting explicitly for portability",
ast_expression::operator_string(op));
}
type_a = value_a->type;
type_b = value_b->type;
}
/* "The fundamental types of the operands (signed or unsigned) must
* match,"
*/
if (type_a->base_type != type_b->base_type) {
_mesa_glsl_error(loc, state, "operands of `%s' must have the same "
"base type", ast_expression::operator_string(op));
return glsl_type::error_type;
}
/* "The operands cannot be vectors of differing size." */
if (type_a->is_vector() &&
type_b->is_vector() &&
type_a->vector_elements != type_b->vector_elements) {
_mesa_glsl_error(loc, state, "operands of `%s' cannot be vectors of "
"different sizes", ast_expression::operator_string(op));
return glsl_type::error_type;
}
/* "If one operand is a scalar and the other a vector, the scalar is
* applied component-wise to the vector, resulting in the same type as
* the vector. The fundamental types of the operands [...] will be the
* resulting fundamental type."
*/
if (type_a->is_scalar())
return type_b;
else
return type_a;
}
static const struct glsl_type *
modulus_result_type(ir_rvalue * &value_a, ir_rvalue * &value_b,
struct _mesa_glsl_parse_state *state, YYLTYPE *loc)
{
const glsl_type *type_a = value_a->type;
const glsl_type *type_b = value_b->type;
if (!state->EXT_gpu_shader4_enable &&
!state->check_version(130, 300, loc, "operator '%%' is reserved")) {
return glsl_type::error_type;
}
/* Section 5.9 (Expressions) of the GLSL 4.00 specification says:
*
* "The operator modulus (%) operates on signed or unsigned integers or
* integer vectors."
*/
if (!type_a->is_integer_32_64()) {
_mesa_glsl_error(loc, state, "LHS of operator %% must be an integer");
return glsl_type::error_type;
}
if (!type_b->is_integer_32_64()) {
_mesa_glsl_error(loc, state, "RHS of operator %% must be an integer");
return glsl_type::error_type;
}
/* "If the fundamental types in the operands do not match, then the
* conversions from section 4.1.10 "Implicit Conversions" are applied
* to create matching types."
*
* Note that GLSL 4.00 (and GL_ARB_gpu_shader5) introduced implicit
* int -> uint conversion rules. Prior to that, there were no implicit
* conversions. So it's harmless to apply them universally - no implicit
* conversions will exist. If the types don't match, we'll receive false,
* and raise an error, satisfying the GLSL 1.50 spec, page 56:
*
* "The operand types must both be signed or unsigned."
*/
if (!apply_implicit_conversion(type_a, value_b, state) &&
!apply_implicit_conversion(type_b, value_a, state)) {
_mesa_glsl_error(loc, state,
"could not implicitly convert operands to "
"modulus (%%) operator");
return glsl_type::error_type;
}
type_a = value_a->type;
type_b = value_b->type;
/* "The operands cannot be vectors of differing size. If one operand is
* a scalar and the other vector, then the scalar is applied component-
* wise to the vector, resulting in the same type as the vector. If both
* are vectors of the same size, the result is computed component-wise."
*/
if (type_a->is_vector()) {
if (!type_b->is_vector()
|| (type_a->vector_elements == type_b->vector_elements))
return type_a;
} else
return type_b;
/* "The operator modulus (%) is not defined for any other data types
* (non-integer types)."
*/
_mesa_glsl_error(loc, state, "type mismatch");
return glsl_type::error_type;
}
static const struct glsl_type *
relational_result_type(ir_rvalue * &value_a, ir_rvalue * &value_b,
struct _mesa_glsl_parse_state *state, YYLTYPE *loc)
{
const glsl_type *type_a = value_a->type;
const glsl_type *type_b = value_b->type;
/* From GLSL 1.50 spec, page 56:
* "The relational operators greater than (>), less than (<), greater
* than or equal (>=), and less than or equal (<=) operate only on
* scalar integer and scalar floating-point expressions."
*/
if (!type_a->is_numeric()
|| !type_b->is_numeric()
|| !type_a->is_scalar()
|| !type_b->is_scalar()) {
_mesa_glsl_error(loc, state,
"operands to relational operators must be scalar and "
"numeric");
return glsl_type::error_type;
}
/* "Either the operands' types must match, or the conversions from
* Section 4.1.10 "Implicit Conversions" will be applied to the integer
* operand, after which the types must match."
*/
if (!apply_implicit_conversion(type_a, value_b, state)
&& !apply_implicit_conversion(type_b, value_a, state)) {
_mesa_glsl_error(loc, state,
"could not implicitly convert operands to "
"relational operator");
return glsl_type::error_type;
}
type_a = value_a->type;
type_b = value_b->type;
if (type_a->base_type != type_b->base_type) {
_mesa_glsl_error(loc, state, "base type mismatch");
return glsl_type::error_type;
}
/* "The result is scalar Boolean."
*/
return glsl_type::bool_type;
}
/**
* \brief Return the result type of a bit-shift operation.
*
* If the given types to the bit-shift operator are invalid, return
* glsl_type::error_type.
*
* \param type_a Type of LHS of bit-shift op
* \param type_b Type of RHS of bit-shift op
*/
static const struct glsl_type *
shift_result_type(const struct glsl_type *type_a,
const struct glsl_type *type_b,
ast_operators op,
struct _mesa_glsl_parse_state *state, YYLTYPE *loc)
{
if (!state->check_bitwise_operations_allowed(loc)) {
return glsl_type::error_type;
}
/* From page 50 (page 56 of the PDF) of the GLSL 1.30 spec:
*
* "The shift operators (<<) and (>>). For both operators, the operands
* must be signed or unsigned integers or integer vectors. One operand
* can be signed while the other is unsigned."
*/
if (!type_a->is_integer_32_64()) {
_mesa_glsl_error(loc, state, "LHS of operator %s must be an integer or "
"integer vector", ast_expression::operator_string(op));
return glsl_type::error_type;
}
if (!type_b->is_integer_32()) {
_mesa_glsl_error(loc, state, "RHS of operator %s must be an integer or "
"integer vector", ast_expression::operator_string(op));
return glsl_type::error_type;
}
/* "If the first operand is a scalar, the second operand has to be
* a scalar as well."
*/
if (type_a->is_scalar() && !type_b->is_scalar()) {
_mesa_glsl_error(loc, state, "if the first operand of %s is scalar, the "
"second must be scalar as well",
ast_expression::operator_string(op));
return glsl_type::error_type;
}
/* If both operands are vectors, check that they have same number of
* elements.
*/
if (type_a->is_vector() &&
type_b->is_vector() &&
type_a->vector_elements != type_b->vector_elements) {
_mesa_glsl_error(loc, state, "vector operands to operator %s must "
"have same number of elements",
ast_expression::operator_string(op));
return glsl_type::error_type;
}
/* "In all cases, the resulting type will be the same type as the left
* operand."
*/
return type_a;
}
/**
* Returns the innermost array index expression in an rvalue tree.
* This is the largest indexing level -- if an array of blocks, then
* it is the block index rather than an indexing expression for an
* array-typed member of an array of blocks.
*/
static ir_rvalue *
find_innermost_array_index(ir_rvalue *rv)
{
ir_dereference_array *last = NULL;
while (rv) {
if (rv->as_dereference_array()) {
last = rv->as_dereference_array();
rv = last->array;
} else if (rv->as_dereference_record())
rv = rv->as_dereference_record()->record;
else if (rv->as_swizzle())
rv = rv->as_swizzle()->val;
else
rv = NULL;
}
if (last)
return last->array_index;
return NULL;
}
/**
* Validates that a value can be assigned to a location with a specified type
*
* Validates that \c rhs can be assigned to some location. If the types are
* not an exact match but an automatic conversion is possible, \c rhs will be
* converted.
*
* \return
* \c NULL if \c rhs cannot be assigned to a location with type \c lhs_type.
* Otherwise the actual RHS to be assigned will be returned. This may be
* \c rhs, or it may be \c rhs after some type conversion.
*
* \note
* In addition to being used for assignments, this function is used to
* type-check return values.
*/
static ir_rvalue *
validate_assignment(struct _mesa_glsl_parse_state *state,
YYLTYPE loc, ir_rvalue *lhs,
ir_rvalue *rhs, bool is_initializer)
{
/* If there is already some error in the RHS, just return it. Anything
* else will lead to an avalanche of error message back to the user.
*/
if (rhs->type->is_error())
return rhs;
/* In the Tessellation Control Shader:
* If a per-vertex output variable is used as an l-value, it is an error
* if the expression indicating the vertex number is not the identifier
* `gl_InvocationID`.
*/
if (state->stage == MESA_SHADER_TESS_CTRL && !lhs->type->is_error()) {
ir_variable *var = lhs->variable_referenced();
if (var && var->data.mode == ir_var_shader_out && !var->data.patch) {
ir_rvalue *index = find_innermost_array_index(lhs);
ir_variable *index_var = index ? index->variable_referenced() : NULL;
if (!index_var || strcmp(index_var->name, "gl_InvocationID") != 0) {
_mesa_glsl_error(&loc, state,
"Tessellation control shader outputs can only "
"be indexed by gl_InvocationID");
return NULL;
}
}
}
/* If the types are identical, the assignment can trivially proceed.
*/
if (rhs->type == lhs->type)
return rhs;
/* If the array element types are the same and the LHS is unsized,
* the assignment is okay for initializers embedded in variable
* declarations.
*
* Note: Whole-array assignments are not permitted in GLSL 1.10, but this
* is handled by ir_dereference::is_lvalue.
*/
const glsl_type *lhs_t = lhs->type;
const glsl_type *rhs_t = rhs->type;
bool unsized_array = false;
while(lhs_t->is_array()) {
if (rhs_t == lhs_t)
break; /* the rest of the inner arrays match so break out early */
if (!rhs_t->is_array()) {
unsized_array = false;
break; /* number of dimensions mismatch */
}
if (lhs_t->length == rhs_t->length) {
lhs_t = lhs_t->fields.array;
rhs_t = rhs_t->fields.array;
continue;
} else if (lhs_t->is_unsized_array()) {
unsized_array = true;
} else {
unsized_array = false;
break; /* sized array mismatch */
}
lhs_t = lhs_t->fields.array;
rhs_t = rhs_t->fields.array;
}
if (unsized_array) {
if (is_initializer) {
if (rhs->type->get_scalar_type() == lhs->type->get_scalar_type())
return rhs;
} else {
_mesa_glsl_error(&loc, state,
"implicitly sized arrays cannot be assigned");
return NULL;
}
}
/* Check for implicit conversion in GLSL 1.20 */
if (apply_implicit_conversion(lhs->type, rhs, state)) {
if (rhs->type == lhs->type)
return rhs;
}
_mesa_glsl_error(&loc, state,
"%s of type %s cannot be assigned to "
"variable of type %s",
is_initializer ? "initializer" : "value",
rhs->type->name, lhs->type->name);
return NULL;
}
static void
mark_whole_array_access(ir_rvalue *access)
{
ir_dereference_variable *deref = access->as_dereference_variable();
if (deref && deref->var) {
deref->var->data.max_array_access = deref->type->length - 1;
}
}
static bool
do_assignment(exec_list *instructions, struct _mesa_glsl_parse_state *state,
const char *non_lvalue_description,
ir_rvalue *lhs, ir_rvalue *rhs,
ir_rvalue **out_rvalue, bool needs_rvalue,
bool is_initializer,
YYLTYPE lhs_loc)
{
void *ctx = state;
bool error_emitted = (lhs->type->is_error() || rhs->type->is_error());
ir_variable *lhs_var = lhs->variable_referenced();
if (lhs_var)
lhs_var->data.assigned = true;
if (!error_emitted) {
if (non_lvalue_description != NULL) {
_mesa_glsl_error(&lhs_loc, state,
"assignment to %s",
non_lvalue_description);
error_emitted = true;
} else if (lhs_var != NULL && (lhs_var->data.read_only ||
(lhs_var->data.mode == ir_var_shader_storage &&
lhs_var->data.memory_read_only))) {
/* We can have memory_read_only set on both images and buffer variables,
* but in the former there is a distinction between assignments to
* the variable itself (read_only) and to the memory they point to
* (memory_read_only), while in the case of buffer variables there is
* no such distinction, that is why this check here is limited to
* buffer variables alone.
*/
_mesa_glsl_error(&lhs_loc, state,
"assignment to read-only variable '%s'",
lhs_var->name);
error_emitted = true;
} else if (lhs->type->is_array() &&
!state->check_version(120, 300, &lhs_loc,
"whole array assignment forbidden")) {
/* From page 32 (page 38 of the PDF) of the GLSL 1.10 spec:
*
* "Other binary or unary expressions, non-dereferenced
* arrays, function names, swizzles with repeated fields,
* and constants cannot be l-values."
*
* The restriction on arrays is lifted in GLSL 1.20 and GLSL ES 3.00.
*/
error_emitted = true;
} else if (!lhs->is_lvalue(state)) {
_mesa_glsl_error(& lhs_loc, state, "non-lvalue in assignment");
error_emitted = true;
}
}
ir_rvalue *new_rhs =
validate_assignment(state, lhs_loc, lhs, rhs, is_initializer);
if (new_rhs != NULL) {
rhs = new_rhs;
/* If the LHS array was not declared with a size, it takes it size from
* the RHS. If the LHS is an l-value and a whole array, it must be a
* dereference of a variable. Any other case would require that the LHS
* is either not an l-value or not a whole array.
*/
if (lhs->type->is_unsized_array()) {
ir_dereference *const d = lhs->as_dereference();
assert(d != NULL);
ir_variable *const var = d->variable_referenced();
assert(var != NULL);
if (var->data.max_array_access >= rhs->type->array_size()) {
/* FINISHME: This should actually log the location of the RHS. */
_mesa_glsl_error(& lhs_loc, state, "array size must be > %u due to "
"previous access",
var->data.max_array_access);
}
var->type = glsl_type::get_array_instance(lhs->type->fields.array,
rhs->type->array_size());
d->type = var->type;
}
if (lhs->type->is_array()) {
mark_whole_array_access(rhs);
mark_whole_array_access(lhs);
}
} else {
error_emitted = true;
}
/* Most callers of do_assignment (assign, add_assign, pre_inc/dec,
* but not post_inc) need the converted assigned value as an rvalue
* to handle things like:
*
* i = j += 1;
*/
if (needs_rvalue) {
ir_rvalue *rvalue;
if (!error_emitted) {
ir_variable *var = new(ctx) ir_variable(rhs->type, "assignment_tmp",
ir_var_temporary);
instructions->push_tail(var);
instructions->push_tail(assign(var, rhs));
ir_dereference_variable *deref_var =
new(ctx) ir_dereference_variable(var);
instructions->push_tail(new(ctx) ir_assignment(lhs, deref_var));
rvalue = new(ctx) ir_dereference_variable(var);
} else {
rvalue = ir_rvalue::error_value(ctx);
}
*out_rvalue = rvalue;
} else {
if (!error_emitted)
instructions->push_tail(new(ctx) ir_assignment(lhs, rhs));
*out_rvalue = NULL;
}
return error_emitted;
}
static ir_rvalue *
get_lvalue_copy(exec_list *instructions, ir_rvalue *lvalue)
{
void *ctx = ralloc_parent(lvalue);
ir_variable *var;
var = new(ctx) ir_variable(lvalue->type, "_post_incdec_tmp",
ir_var_temporary);
instructions->push_tail(var);
instructions->push_tail(new(ctx) ir_assignment(new(ctx) ir_dereference_variable(var),
lvalue));
return new(ctx) ir_dereference_variable(var);
}
ir_rvalue *
ast_node::hir(exec_list *instructions, struct _mesa_glsl_parse_state *state)
{
(void) instructions;
(void) state;
return NULL;
}
bool
ast_node::has_sequence_subexpression() const
{
return false;
}
void
ast_node::set_is_lhs(bool /* new_value */)
{
}
void
ast_function_expression::hir_no_rvalue(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
(void)hir(instructions, state);
}
void
ast_aggregate_initializer::hir_no_rvalue(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
(void)hir(instructions, state);
}
static ir_rvalue *
do_comparison(void *mem_ctx, int operation, ir_rvalue *op0, ir_rvalue *op1)
{
int join_op;
ir_rvalue *cmp = NULL;
if (operation == ir_binop_all_equal)
join_op = ir_binop_logic_and;
else
join_op = ir_binop_logic_or;
switch (op0->type->base_type) {
case GLSL_TYPE_FLOAT:
case GLSL_TYPE_FLOAT16:
case GLSL_TYPE_UINT:
case GLSL_TYPE_INT:
case GLSL_TYPE_BOOL:
case GLSL_TYPE_DOUBLE:
case GLSL_TYPE_UINT64:
case GLSL_TYPE_INT64:
case GLSL_TYPE_UINT16:
case GLSL_TYPE_INT16:
case GLSL_TYPE_UINT8:
case GLSL_TYPE_INT8:
return new(mem_ctx) ir_expression(operation, op0, op1);
case GLSL_TYPE_ARRAY: {
for (unsigned int i = 0; i < op0->type->length; i++) {
ir_rvalue *e0, *e1, *result;
e0 = new(mem_ctx) ir_dereference_array(op0->clone(mem_ctx, NULL),
new(mem_ctx) ir_constant(i));
e1 = new(mem_ctx) ir_dereference_array(op1->clone(mem_ctx, NULL),
new(mem_ctx) ir_constant(i));
result = do_comparison(mem_ctx, operation, e0, e1);
if (cmp) {
cmp = new(mem_ctx) ir_expression(join_op, cmp, result);
} else {
cmp = result;
}
}
mark_whole_array_access(op0);
mark_whole_array_access(op1);
break;
}
case GLSL_TYPE_STRUCT: {
for (unsigned int i = 0; i < op0->type->length; i++) {
ir_rvalue *e0, *e1, *result;
const char *field_name = op0->type->fields.structure[i].name;
e0 = new(mem_ctx) ir_dereference_record(op0->clone(mem_ctx, NULL),
field_name);
e1 = new(mem_ctx) ir_dereference_record(op1->clone(mem_ctx, NULL),
field_name);
result = do_comparison(mem_ctx, operation, e0, e1);
if (cmp) {
cmp = new(mem_ctx) ir_expression(join_op, cmp, result);
} else {
cmp = result;
}
}
break;
}
case GLSL_TYPE_ERROR:
case GLSL_TYPE_VOID:
case GLSL_TYPE_SAMPLER:
case GLSL_TYPE_IMAGE:
case GLSL_TYPE_INTERFACE:
case GLSL_TYPE_ATOMIC_UINT:
case GLSL_TYPE_SUBROUTINE:
case GLSL_TYPE_FUNCTION:
/* I assume a comparison of a struct containing a sampler just
* ignores the sampler present in the type.
*/
break;
}
if (cmp == NULL)
cmp = new(mem_ctx) ir_constant(true);
return cmp;
}
/* For logical operations, we want to ensure that the operands are
* scalar booleans. If it isn't, emit an error and return a constant
* boolean to avoid triggering cascading error messages.
*/
static ir_rvalue *
get_scalar_boolean_operand(exec_list *instructions,
struct _mesa_glsl_parse_state *state,
ast_expression *parent_expr,
int operand,
const char *operand_name,
bool *error_emitted)
{
ast_expression *expr = parent_expr->subexpressions[operand];
void *ctx = state;
ir_rvalue *val = expr->hir(instructions, state);
if (val->type->is_boolean() && val->type->is_scalar())
return val;
if (!*error_emitted) {
YYLTYPE loc = expr->get_location();
_mesa_glsl_error(&loc, state, "%s of `%s' must be scalar boolean",
operand_name,
parent_expr->operator_string(parent_expr->oper));
*error_emitted = true;
}
return new(ctx) ir_constant(true);
}
/**
* If name refers to a builtin array whose maximum allowed size is less than
* size, report an error and return true. Otherwise return false.
*/
void
check_builtin_array_max_size(const char *name, unsigned size,
YYLTYPE loc, struct _mesa_glsl_parse_state *state)
{
if ((strcmp("gl_TexCoord", name) == 0)
&& (size > state->Const.MaxTextureCoords)) {
/* From page 54 (page 60 of the PDF) of the GLSL 1.20 spec:
*
* "The size [of gl_TexCoord] can be at most
* gl_MaxTextureCoords."
*/
_mesa_glsl_error(&loc, state, "`gl_TexCoord' array size cannot "
"be larger than gl_MaxTextureCoords (%u)",
state->Const.MaxTextureCoords);
} else if (strcmp("gl_ClipDistance", name) == 0) {
state->clip_dist_size = size;
if (size + state->cull_dist_size > state->Const.MaxClipPlanes) {
/* From section 7.1 (Vertex Shader Special Variables) of the
* GLSL 1.30 spec:
*
* "The gl_ClipDistance array is predeclared as unsized and
* must be sized by the shader either redeclaring it with a
* size or indexing it only with integral constant
* expressions. ... The size can be at most
* gl_MaxClipDistances."
*/
_mesa_glsl_error(&loc, state, "`gl_ClipDistance' array size cannot "
"be larger than gl_MaxClipDistances (%u)",
state->Const.MaxClipPlanes);
}
} else if (strcmp("gl_CullDistance", name) == 0) {
state->cull_dist_size = size;
if (size + state->clip_dist_size > state->Const.MaxClipPlanes) {
/* From the ARB_cull_distance spec:
*
* "The gl_CullDistance array is predeclared as unsized and
* must be sized by the shader either redeclaring it with
* a size or indexing it only with integral constant
* expressions. The size determines the number and set of
* enabled cull distances and can be at most
* gl_MaxCullDistances."
*/
_mesa_glsl_error(&loc, state, "`gl_CullDistance' array size cannot "
"be larger than gl_MaxCullDistances (%u)",
state->Const.MaxClipPlanes);
}
}
}
/**
* Create the constant 1, of a which is appropriate for incrementing and
* decrementing values of the given GLSL type. For example, if type is vec4,
* this creates a constant value of 1.0 having type float.
*
* If the given type is invalid for increment and decrement operators, return
* a floating point 1--the error will be detected later.
*/
static ir_rvalue *
constant_one_for_inc_dec(void *ctx, const glsl_type *type)
{
switch (type->base_type) {
case GLSL_TYPE_UINT:
return new(ctx) ir_constant((unsigned) 1);
case GLSL_TYPE_INT:
return new(ctx) ir_constant(1);
case GLSL_TYPE_UINT64:
return new(ctx) ir_constant((uint64_t) 1);
case GLSL_TYPE_INT64:
return new(ctx) ir_constant((int64_t) 1);
default:
case GLSL_TYPE_FLOAT:
return new(ctx) ir_constant(1.0f);
}
}
ir_rvalue *
ast_expression::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
return do_hir(instructions, state, true);
}
void
ast_expression::hir_no_rvalue(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
do_hir(instructions, state, false);
}
void
ast_expression::set_is_lhs(bool new_value)
{
/* is_lhs is tracked only to print "variable used uninitialized" warnings,
* if we lack an identifier we can just skip it.
*/
if (this->primary_expression.identifier == NULL)
return;
this->is_lhs = new_value;
/* We need to go through the subexpressions tree to cover cases like
* ast_field_selection
*/
if (this->subexpressions[0] != NULL)
this->subexpressions[0]->set_is_lhs(new_value);
}
ir_rvalue *
ast_expression::do_hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state,
bool needs_rvalue)
{
void *ctx = state;
static const int operations[AST_NUM_OPERATORS] = {
-1, /* ast_assign doesn't convert to ir_expression. */
-1, /* ast_plus doesn't convert to ir_expression. */
ir_unop_neg,
ir_binop_add,
ir_binop_sub,
ir_binop_mul,
ir_binop_div,
ir_binop_mod,
ir_binop_lshift,
ir_binop_rshift,
ir_binop_less,
ir_binop_less, /* This is correct. See the ast_greater case below. */
ir_binop_gequal, /* This is correct. See the ast_lequal case below. */
ir_binop_gequal,
ir_binop_all_equal,
ir_binop_any_nequal,
ir_binop_bit_and,
ir_binop_bit_xor,
ir_binop_bit_or,
ir_unop_bit_not,
ir_binop_logic_and,
ir_binop_logic_xor,
ir_binop_logic_or,
ir_unop_logic_not,
/* Note: The following block of expression types actually convert
* to multiple IR instructions.
*/
ir_binop_mul, /* ast_mul_assign */
ir_binop_div, /* ast_div_assign */
ir_binop_mod, /* ast_mod_assign */
ir_binop_add, /* ast_add_assign */
ir_binop_sub, /* ast_sub_assign */
ir_binop_lshift, /* ast_ls_assign */
ir_binop_rshift, /* ast_rs_assign */
ir_binop_bit_and, /* ast_and_assign */
ir_binop_bit_xor, /* ast_xor_assign */
ir_binop_bit_or, /* ast_or_assign */
-1, /* ast_conditional doesn't convert to ir_expression. */
ir_binop_add, /* ast_pre_inc. */
ir_binop_sub, /* ast_pre_dec. */
ir_binop_add, /* ast_post_inc. */
ir_binop_sub, /* ast_post_dec. */
-1, /* ast_field_selection doesn't conv to ir_expression. */
-1, /* ast_array_index doesn't convert to ir_expression. */
-1, /* ast_function_call doesn't conv to ir_expression. */
-1, /* ast_identifier doesn't convert to ir_expression. */
-1, /* ast_int_constant doesn't convert to ir_expression. */
-1, /* ast_uint_constant doesn't conv to ir_expression. */
-1, /* ast_float_constant doesn't conv to ir_expression. */
-1, /* ast_bool_constant doesn't conv to ir_expression. */
-1, /* ast_sequence doesn't convert to ir_expression. */
-1, /* ast_aggregate shouldn't ever even get here. */
};
ir_rvalue *result = NULL;
ir_rvalue *op[3];
const struct glsl_type *type, *orig_type;
bool error_emitted = false;
YYLTYPE loc;
loc = this->get_location();
switch (this->oper) {
case ast_aggregate:
unreachable("ast_aggregate: Should never get here.");
case ast_assign: {
this->subexpressions[0]->set_is_lhs(true);
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
error_emitted =
do_assignment(instructions, state,
this->subexpressions[0]->non_lvalue_description,
op[0], op[1], &result, needs_rvalue, false,
this->subexpressions[0]->get_location());
break;
}
case ast_plus:
op[0] = this->subexpressions[0]->hir(instructions, state);
type = unary_arithmetic_result_type(op[0]->type, state, & loc);
error_emitted = type->is_error();
result = op[0];
break;
case ast_neg:
op[0] = this->subexpressions[0]->hir(instructions, state);
type = unary_arithmetic_result_type(op[0]->type, state, & loc);
error_emitted = type->is_error();
result = new(ctx) ir_expression(operations[this->oper], type,
op[0], NULL);
break;
case ast_add:
case ast_sub:
case ast_mul:
case ast_div:
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
type = arithmetic_result_type(op[0], op[1],
(this->oper == ast_mul),
state, & loc);
error_emitted = type->is_error();
result = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
break;
case ast_mod:
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
type = modulus_result_type(op[0], op[1], state, &loc);
assert(operations[this->oper] == ir_binop_mod);
result = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
error_emitted = type->is_error();
break;
case ast_lshift:
case ast_rshift:
if (!state->check_bitwise_operations_allowed(&loc)) {
error_emitted = true;
}
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
type = shift_result_type(op[0]->type, op[1]->type, this->oper, state,
&loc);
result = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
error_emitted = op[0]->type->is_error() || op[1]->type->is_error();
break;
case ast_less:
case ast_greater:
case ast_lequal:
case ast_gequal:
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
type = relational_result_type(op[0], op[1], state, & loc);
/* The relational operators must either generate an error or result
* in a scalar boolean. See page 57 of the GLSL 1.50 spec.
*/
assert(type->is_error()
|| (type->is_boolean() && type->is_scalar()));
/* Like NIR, GLSL IR does not have opcodes for > or <=. Instead, swap
* the arguments and use < or >=.
*/
if (this->oper == ast_greater || this->oper == ast_lequal) {
ir_rvalue *const tmp = op[0];
op[0] = op[1];
op[1] = tmp;
}
result = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
error_emitted = type->is_error();
break;
case ast_nequal:
case ast_equal:
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
/* From page 58 (page 64 of the PDF) of the GLSL 1.50 spec:
*
* "The equality operators equal (==), and not equal (!=)
* operate on all types. They result in a scalar Boolean. If
* the operand types do not match, then there must be a
* conversion from Section 4.1.10 "Implicit Conversions"
* applied to one operand that can make them match, in which
* case this conversion is done."
*/
if (op[0]->type == glsl_type::void_type || op[1]->type == glsl_type::void_type) {
_mesa_glsl_error(& loc, state, "`%s': wrong operand types: "
"no operation `%1$s' exists that takes a left-hand "
"operand of type 'void' or a right operand of type "
"'void'", (this->oper == ast_equal) ? "==" : "!=");
error_emitted = true;
} else if ((!apply_implicit_conversion(op[0]->type, op[1], state)
&& !apply_implicit_conversion(op[1]->type, op[0], state))
|| (op[0]->type != op[1]->type)) {
_mesa_glsl_error(& loc, state, "operands of `%s' must have the same "
"type", (this->oper == ast_equal) ? "==" : "!=");
error_emitted = true;
} else if ((op[0]->type->is_array() || op[1]->type->is_array()) &&
!state->check_version(120, 300, &loc,
"array comparisons forbidden")) {
error_emitted = true;
} else if ((op[0]->type->contains_subroutine() ||
op[1]->type->contains_subroutine())) {
_mesa_glsl_error(&loc, state, "subroutine comparisons forbidden");
error_emitted = true;
} else if ((op[0]->type->contains_opaque() ||
op[1]->type->contains_opaque())) {
_mesa_glsl_error(&loc, state, "opaque type comparisons forbidden");
error_emitted = true;
}
if (error_emitted) {
result = new(ctx) ir_constant(false);
} else {
result = do_comparison(ctx, operations[this->oper], op[0], op[1]);
assert(result->type == glsl_type::bool_type);
}
break;
case ast_bit_and:
case ast_bit_xor:
case ast_bit_or:
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
type = bit_logic_result_type(op[0], op[1], this->oper, state, &loc);
result = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
error_emitted = op[0]->type->is_error() || op[1]->type->is_error();
break;
case ast_bit_not:
op[0] = this->subexpressions[0]->hir(instructions, state);
if (!state->check_bitwise_operations_allowed(&loc)) {
error_emitted = true;
}
if (!op[0]->type->is_integer_32_64()) {
_mesa_glsl_error(&loc, state, "operand of `~' must be an integer");
error_emitted = true;
}
type = error_emitted ? glsl_type::error_type : op[0]->type;
result = new(ctx) ir_expression(ir_unop_bit_not, type, op[0], NULL);
break;
case ast_logic_and: {
exec_list rhs_instructions;
op[0] = get_scalar_boolean_operand(instructions, state, this, 0,
"LHS", &error_emitted);
op[1] = get_scalar_boolean_operand(&rhs_instructions, state, this, 1,
"RHS", &error_emitted);
if (rhs_instructions.is_empty()) {
result = new(ctx) ir_expression(ir_binop_logic_and, op[0], op[1]);
} else {
ir_variable *const tmp = new(ctx) ir_variable(glsl_type::bool_type,
"and_tmp",
ir_var_temporary);
instructions->push_tail(tmp);
ir_if *const stmt = new(ctx) ir_if(op[0]);
instructions->push_tail(stmt);
stmt->then_instructions.append_list(&rhs_instructions);
ir_dereference *const then_deref = new(ctx) ir_dereference_variable(tmp);
ir_assignment *const then_assign =
new(ctx) ir_assignment(then_deref, op[1]);
stmt->then_instructions.push_tail(then_assign);
ir_dereference *const else_deref = new(ctx) ir_dereference_variable(tmp);
ir_assignment *const else_assign =
new(ctx) ir_assignment(else_deref, new(ctx) ir_constant(false));
stmt->else_instructions.push_tail(else_assign);
result = new(ctx) ir_dereference_variable(tmp);
}
break;
}
case ast_logic_or: {
exec_list rhs_instructions;
op[0] = get_scalar_boolean_operand(instructions, state, this, 0,
"LHS", &error_emitted);
op[1] = get_scalar_boolean_operand(&rhs_instructions, state, this, 1,
"RHS", &error_emitted);
if (rhs_instructions.is_empty()) {
result = new(ctx) ir_expression(ir_binop_logic_or, op[0], op[1]);
} else {
ir_variable *const tmp = new(ctx) ir_variable(glsl_type::bool_type,
"or_tmp",
ir_var_temporary);
instructions->push_tail(tmp);
ir_if *const stmt = new(ctx) ir_if(op[0]);
instructions->push_tail(stmt);
ir_dereference *const then_deref = new(ctx) ir_dereference_variable(tmp);
ir_assignment *const then_assign =
new(ctx) ir_assignment(then_deref, new(ctx) ir_constant(true));
stmt->then_instructions.push_tail(then_assign);
stmt->else_instructions.append_list(&rhs_instructions);
ir_dereference *const else_deref = new(ctx) ir_dereference_variable(tmp);
ir_assignment *const else_assign =
new(ctx) ir_assignment(else_deref, op[1]);
stmt->else_instructions.push_tail(else_assign);
result = new(ctx) ir_dereference_variable(tmp);
}
break;
}
case ast_logic_xor:
/* From page 33 (page 39 of the PDF) of the GLSL 1.10 spec:
*
* "The logical binary operators and (&&), or ( | | ), and
* exclusive or (^^). They operate only on two Boolean
* expressions and result in a Boolean expression."
*/
op[0] = get_scalar_boolean_operand(instructions, state, this, 0, "LHS",
&error_emitted);
op[1] = get_scalar_boolean_operand(instructions, state, this, 1, "RHS",
&error_emitted);
result = new(ctx) ir_expression(operations[this->oper], glsl_type::bool_type,
op[0], op[1]);
break;
case ast_logic_not:
op[0] = get_scalar_boolean_operand(instructions, state, this, 0,
"operand", &error_emitted);
result = new(ctx) ir_expression(operations[this->oper], glsl_type::bool_type,
op[0], NULL);
break;
case ast_mul_assign:
case ast_div_assign:
case ast_add_assign:
case ast_sub_assign: {
this->subexpressions[0]->set_is_lhs(true);
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
orig_type = op[0]->type;
/* Break out if operand types were not parsed successfully. */
if ((op[0]->type == glsl_type::error_type ||
op[1]->type == glsl_type::error_type)) {
error_emitted = true;
break;
}
type = arithmetic_result_type(op[0], op[1],
(this->oper == ast_mul_assign),
state, & loc);
if (type != orig_type) {
_mesa_glsl_error(& loc, state,
"could not implicitly convert "
"%s to %s", type->name, orig_type->name);
type = glsl_type::error_type;
}
ir_rvalue *temp_rhs = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
error_emitted =
do_assignment(instructions, state,
this->subexpressions[0]->non_lvalue_description,
op[0]->clone(ctx, NULL), temp_rhs,
&result, needs_rvalue, false,
this->subexpressions[0]->get_location());
/* GLSL 1.10 does not allow array assignment. However, we don't have to
* explicitly test for this because none of the binary expression
* operators allow array operands either.
*/
break;
}
case ast_mod_assign: {
this->subexpressions[0]->set_is_lhs(true);
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
orig_type = op[0]->type;
type = modulus_result_type(op[0], op[1], state, &loc);
if (type != orig_type) {
_mesa_glsl_error(& loc, state,
"could not implicitly convert "
"%s to %s", type->name, orig_type->name);
type = glsl_type::error_type;
}
assert(operations[this->oper] == ir_binop_mod);
ir_rvalue *temp_rhs;
temp_rhs = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
error_emitted =
do_assignment(instructions, state,
this->subexpressions[0]->non_lvalue_description,
op[0]->clone(ctx, NULL), temp_rhs,
&result, needs_rvalue, false,
this->subexpressions[0]->get_location());
break;
}
case ast_ls_assign:
case ast_rs_assign: {
this->subexpressions[0]->set_is_lhs(true);
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
type = shift_result_type(op[0]->type, op[1]->type, this->oper, state,
&loc);
ir_rvalue *temp_rhs = new(ctx) ir_expression(operations[this->oper],
type, op[0], op[1]);
error_emitted =
do_assignment(instructions, state,
this->subexpressions[0]->non_lvalue_description,
op[0]->clone(ctx, NULL), temp_rhs,
&result, needs_rvalue, false,
this->subexpressions[0]->get_location());
break;
}
case ast_and_assign:
case ast_xor_assign:
case ast_or_assign: {
this->subexpressions[0]->set_is_lhs(true);
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = this->subexpressions[1]->hir(instructions, state);
orig_type = op[0]->type;
type = bit_logic_result_type(op[0], op[1], this->oper, state, &loc);
if (type != orig_type) {
_mesa_glsl_error(& loc, state,
"could not implicitly convert "
"%s to %s", type->name, orig_type->name);
type = glsl_type::error_type;
}
ir_rvalue *temp_rhs = new(ctx) ir_expression(operations[this->oper],
type, op[0], op[1]);
error_emitted =
do_assignment(instructions, state,
this->subexpressions[0]->non_lvalue_description,
op[0]->clone(ctx, NULL), temp_rhs,
&result, needs_rvalue, false,
this->subexpressions[0]->get_location());
break;
}
case ast_conditional: {
/* From page 59 (page 65 of the PDF) of the GLSL 1.50 spec:
*
* "The ternary selection operator (?:). It operates on three
* expressions (exp1 ? exp2 : exp3). This operator evaluates the
* first expression, which must result in a scalar Boolean."
*/
op[0] = get_scalar_boolean_operand(instructions, state, this, 0,
"condition", &error_emitted);
/* The :? operator is implemented by generating an anonymous temporary
* followed by an if-statement. The last instruction in each branch of
* the if-statement assigns a value to the anonymous temporary. This
* temporary is the r-value of the expression.
*/
exec_list then_instructions;
exec_list else_instructions;
op[1] = this->subexpressions[1]->hir(&then_instructions, state);
op[2] = this->subexpressions[2]->hir(&else_instructions, state);
/* From page 59 (page 65 of the PDF) of the GLSL 1.50 spec:
*
* "The second and third expressions can be any type, as
* long their types match, or there is a conversion in
* Section 4.1.10 "Implicit Conversions" that can be applied
* to one of the expressions to make their types match. This
* resulting matching type is the type of the entire
* expression."
*/
if ((!apply_implicit_conversion(op[1]->type, op[2], state)
&& !apply_implicit_conversion(op[2]->type, op[1], state))
|| (op[1]->type != op[2]->type)) {
YYLTYPE loc = this->subexpressions[1]->get_location();
_mesa_glsl_error(& loc, state, "second and third operands of ?: "
"operator must have matching types");
error_emitted = true;
type = glsl_type::error_type;
} else {
type = op[1]->type;
}
/* From page 33 (page 39 of the PDF) of the GLSL 1.10 spec:
*
* "The second and third expressions must be the same type, but can
* be of any type other than an array."
*/
if (type->is_array() &&
!state->check_version(120, 300, &loc,
"second and third operands of ?: operator "
"cannot be arrays")) {
error_emitted = true;
}
/* From section 4.1.7 of the GLSL 4.50 spec (Opaque Types):
*
* "Except for array indexing, structure member selection, and
* parentheses, opaque variables are not allowed to be operands in
* expressions; such use results in a compile-time error."
*/
if (type->contains_opaque()) {
if (!(state->has_bindless() && (type->is_image() || type->is_sampler()))) {
_mesa_glsl_error(&loc, state, "variables of type %s cannot be "
"operands of the ?: operator", type->name);
error_emitted = true;
}
}
ir_constant *cond_val = op[0]->constant_expression_value(ctx);
if (then_instructions.is_empty()
&& else_instructions.is_empty()
&& cond_val != NULL) {
result = cond_val->value.b[0] ? op[1] : op[2];
} else {
/* The copy to conditional_tmp reads the whole array. */
if (type->is_array()) {
mark_whole_array_access(op[1]);
mark_whole_array_access(op[2]);
}
ir_variable *const tmp =
new(ctx) ir_variable(type, "conditional_tmp", ir_var_temporary);
instructions->push_tail(tmp);
ir_if *const stmt = new(ctx) ir_if(op[0]);
instructions->push_tail(stmt);
then_instructions.move_nodes_to(& stmt->then_instructions);
ir_dereference *const then_deref =
new(ctx) ir_dereference_variable(tmp);
ir_assignment *const then_assign =
new(ctx) ir_assignment(then_deref, op[1]);
stmt->then_instructions.push_tail(then_assign);
else_instructions.move_nodes_to(& stmt->else_instructions);
ir_dereference *const else_deref =
new(ctx) ir_dereference_variable(tmp);
ir_assignment *const else_assign =
new(ctx) ir_assignment(else_deref, op[2]);
stmt->else_instructions.push_tail(else_assign);
result = new(ctx) ir_dereference_variable(tmp);
}
break;
}
case ast_pre_inc:
case ast_pre_dec: {
this->non_lvalue_description = (this->oper == ast_pre_inc)
? "pre-increment operation" : "pre-decrement operation";
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = constant_one_for_inc_dec(ctx, op[0]->type);
type = arithmetic_result_type(op[0], op[1], false, state, & loc);
ir_rvalue *temp_rhs;
temp_rhs = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
error_emitted =
do_assignment(instructions, state,
this->subexpressions[0]->non_lvalue_description,
op[0]->clone(ctx, NULL), temp_rhs,
&result, needs_rvalue, false,
this->subexpressions[0]->get_location());
break;
}
case ast_post_inc:
case ast_post_dec: {
this->non_lvalue_description = (this->oper == ast_post_inc)
? "post-increment operation" : "post-decrement operation";
op[0] = this->subexpressions[0]->hir(instructions, state);
op[1] = constant_one_for_inc_dec(ctx, op[0]->type);
error_emitted = op[0]->type->is_error() || op[1]->type->is_error();
if (error_emitted) {
result = ir_rvalue::error_value(ctx);
break;
}
type = arithmetic_result_type(op[0], op[1], false, state, & loc);
ir_rvalue *temp_rhs;
temp_rhs = new(ctx) ir_expression(operations[this->oper], type,
op[0], op[1]);
/* Get a temporary of a copy of the lvalue before it's modified.
* This may get thrown away later.
*/
result = get_lvalue_copy(instructions, op[0]->clone(ctx, NULL));
ir_rvalue *junk_rvalue;
error_emitted =
do_assignment(instructions, state,
this->subexpressions[0]->non_lvalue_description,
op[0]->clone(ctx, NULL), temp_rhs,
&junk_rvalue, false, false,
this->subexpressions[0]->get_location());
break;
}
case ast_field_selection:
result = _mesa_ast_field_selection_to_hir(this, instructions, state);
break;
case ast_array_index: {
YYLTYPE index_loc = subexpressions[1]->get_location();
/* Getting if an array is being used uninitialized is beyond what we get
* from ir_value.data.assigned. Setting is_lhs as true would force to
* not raise a uninitialized warning when using an array
*/
subexpressions[0]->set_is_lhs(true);
op[0] = subexpressions[0]->hir(instructions, state);
op[1] = subexpressions[1]->hir(instructions, state);
result = _mesa_ast_array_index_to_hir(ctx, state, op[0], op[1],
loc, index_loc);
if (result->type->is_error())
error_emitted = true;
break;
}
case ast_unsized_array_dim:
unreachable("ast_unsized_array_dim: Should never get here.");
case ast_function_call:
/* Should *NEVER* get here. ast_function_call should always be handled
* by ast_function_expression::hir.
*/
unreachable("ast_function_call: handled elsewhere ");
case ast_identifier: {
/* ast_identifier can appear several places in a full abstract syntax
* tree. This particular use must be at location specified in the grammar
* as 'variable_identifier'.
*/
ir_variable *var =
state->symbols->get_variable(this->primary_expression.identifier);
if (var == NULL) {
/* the identifier might be a subroutine name */
char *sub_name;
sub_name = ralloc_asprintf(ctx, "%s_%s", _mesa_shader_stage_to_subroutine_prefix(state->stage), this->primary_expression.identifier);
var = state->symbols->get_variable(sub_name);
ralloc_free(sub_name);
}
if (var != NULL) {
var->data.used = true;
result = new(ctx) ir_dereference_variable(var);
if ((var->data.mode == ir_var_auto || var->data.mode == ir_var_shader_out)
&& !this->is_lhs
&& result->variable_referenced()->data.assigned != true
&& !is_gl_identifier(var->name)) {
_mesa_glsl_warning(&loc, state, "`%s' used uninitialized",
this->primary_expression.identifier);
}
/* From the EXT_shader_framebuffer_fetch spec:
*
* "Unless the GL_EXT_shader_framebuffer_fetch extension has been
* enabled in addition, it's an error to use gl_LastFragData if it
* hasn't been explicitly redeclared with layout(noncoherent)."
*/
if (var->data.fb_fetch_output && var->data.memory_coherent &&
!state->EXT_shader_framebuffer_fetch_enable) {
_mesa_glsl_error(&loc, state,
"invalid use of framebuffer fetch output not "
"qualified with layout(noncoherent)");
}
} else {
_mesa_glsl_error(& loc, state, "`%s' undeclared",
this->primary_expression.identifier);
result = ir_rvalue::error_value(ctx);
error_emitted = true;
}
break;
}
case ast_int_constant:
result = new(ctx) ir_constant(this->primary_expression.int_constant);
break;
case ast_uint_constant:
result = new(ctx) ir_constant(this->primary_expression.uint_constant);
break;
case ast_float_constant:
result = new(ctx) ir_constant(this->primary_expression.float_constant);
break;
case ast_bool_constant:
result = new(ctx) ir_constant(bool(this->primary_expression.bool_constant));
break;
case ast_double_constant:
result = new(ctx) ir_constant(this->primary_expression.double_constant);
break;
case ast_uint64_constant:
result = new(ctx) ir_constant(this->primary_expression.uint64_constant);
break;
case ast_int64_constant:
result = new(ctx) ir_constant(this->primary_expression.int64_constant);
break;
case ast_sequence: {
/* It should not be possible to generate a sequence in the AST without
* any expressions in it.
*/
assert(!this->expressions.is_empty());
/* The r-value of a sequence is the last expression in the sequence. If
* the other expressions in the sequence do not have side-effects (and
* therefore add instructions to the instruction list), they get dropped
* on the floor.
*/
exec_node *previous_tail = NULL;
YYLTYPE previous_operand_loc = loc;
foreach_list_typed (ast_node, ast, link, &this->expressions) {
/* If one of the operands of comma operator does not generate any
* code, we want to emit a warning. At each pass through the loop
* previous_tail will point to the last instruction in the stream
* *before* processing the previous operand. Naturally,
* instructions->get_tail_raw() will point to the last instruction in
* the stream *after* processing the previous operand. If the two
* pointers match, then the previous operand had no effect.
*
* The warning behavior here differs slightly from GCC. GCC will
* only emit a warning if none of the left-hand operands have an
* effect. However, it will emit a warning for each. I believe that
* there are some cases in C (especially with GCC extensions) where
* it is useful to have an intermediate step in a sequence have no
* effect, but I don't think these cases exist in GLSL. Either way,
* it would be a giant hassle to replicate that behavior.
*/
if (previous_tail == instructions->get_tail_raw()) {
_mesa_glsl_warning(&previous_operand_loc, state,
"left-hand operand of comma expression has "
"no effect");
}
/* The tail is directly accessed instead of using the get_tail()
* method for performance reasons. get_tail() has extra code to
* return NULL when the list is empty. We don't care about that
* here, so using get_tail_raw() is fine.
*/
previous_tail = instructions->get_tail_raw();
previous_operand_loc = ast->get_location();
result = ast->hir(instructions, state);
}
/* Any errors should have already been emitted in the loop above.
*/
error_emitted = true;
break;
}
}
type = NULL; /* use result->type, not type. */
assert(error_emitted || (result != NULL || !needs_rvalue));
if (result && result->type->is_error() && !error_emitted)
_mesa_glsl_error(& loc, state, "type mismatch");
return result;
}
bool
ast_expression::has_sequence_subexpression() const
{
switch (this->oper) {
case ast_plus:
case ast_neg:
case ast_bit_not:
case ast_logic_not:
case ast_pre_inc:
case ast_pre_dec:
case ast_post_inc:
case ast_post_dec:
return this->subexpressions[0]->has_sequence_subexpression();
case ast_assign:
case ast_add:
case ast_sub:
case ast_mul:
case ast_div:
case ast_mod:
case ast_lshift:
case ast_rshift:
case ast_less:
case ast_greater:
case ast_lequal:
case ast_gequal:
case ast_nequal:
case ast_equal:
case ast_bit_and:
case ast_bit_xor:
case ast_bit_or:
case ast_logic_and:
case ast_logic_or:
case ast_logic_xor:
case ast_array_index:
case ast_mul_assign:
case ast_div_assign:
case ast_add_assign:
case ast_sub_assign:
case ast_mod_assign:
case ast_ls_assign:
case ast_rs_assign:
case ast_and_assign:
case ast_xor_assign:
case ast_or_assign:
return this->subexpressions[0]->has_sequence_subexpression() ||
this->subexpressions[1]->has_sequence_subexpression();
case ast_conditional:
return this->subexpressions[0]->has_sequence_subexpression() ||
this->subexpressions[1]->has_sequence_subexpression() ||
this->subexpressions[2]->has_sequence_subexpression();
case ast_sequence:
return true;
case ast_field_selection:
case ast_identifier:
case ast_int_constant:
case ast_uint_constant:
case ast_float_constant:
case ast_bool_constant:
case ast_double_constant:
case ast_int64_constant:
case ast_uint64_constant:
return false;
case ast_aggregate:
return false;
case ast_function_call:
unreachable("should be handled by ast_function_expression::hir");
case ast_unsized_array_dim:
unreachable("ast_unsized_array_dim: Should never get here.");
}
return false;
}
ir_rvalue *
ast_expression_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
/* It is possible to have expression statements that don't have an
* expression. This is the solitary semicolon:
*
* for (i = 0; i < 5; i++)
* ;
*
* In this case the expression will be NULL. Test for NULL and don't do
* anything in that case.
*/
if (expression != NULL)
expression->hir_no_rvalue(instructions, state);
/* Statements do not have r-values.
*/
return NULL;
}
ir_rvalue *
ast_compound_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
if (new_scope)
state->symbols->push_scope();
foreach_list_typed (ast_node, ast, link, &this->statements)
ast->hir(instructions, state);
if (new_scope)
state->symbols->pop_scope();
/* Compound statements do not have r-values.
*/
return NULL;
}
/**
* Evaluate the given exec_node (which should be an ast_node representing
* a single array dimension) and return its integer value.
*/
static unsigned
process_array_size(exec_node *node,
struct _mesa_glsl_parse_state *state)
{
void *mem_ctx = state;
exec_list dummy_instructions;
ast_node *array_size = exec_node_data(ast_node, node, link);
/**
* Dimensions other than the outermost dimension can by unsized if they
* are immediately sized by a constructor or initializer.
*/
if (((ast_expression*)array_size)->oper == ast_unsized_array_dim)
return 0;
ir_rvalue *const ir = array_size->hir(& dummy_instructions, state);
YYLTYPE loc = array_size->get_location();
if (ir == NULL) {
_mesa_glsl_error(& loc, state,
"array size could not be resolved");
return 0;
}
if (!ir->type->is_integer_32()) {
_mesa_glsl_error(& loc, state,
"array size must be integer type");
return 0;
}
if (!ir->type->is_scalar()) {
_mesa_glsl_error(& loc, state,
"array size must be scalar type");
return 0;
}
ir_constant *const size = ir->constant_expression_value(mem_ctx);
if (size == NULL ||
(state->is_version(120, 300) &&
array_size->has_sequence_subexpression())) {
_mesa_glsl_error(& loc, state, "array size must be a "
"constant valued expression");
return 0;
}
if (size->value.i[0] <= 0) {
_mesa_glsl_error(& loc, state, "array size must be > 0");
return 0;
}
assert(size->type == ir->type);
/* If the array size is const (and we've verified that
* it is) then no instructions should have been emitted
* when we converted it to HIR. If they were emitted,
* then either the array size isn't const after all, or
* we are emitting unnecessary instructions.
*/
assert(dummy_instructions.is_empty());
return size->value.u[0];
}
static const glsl_type *
process_array_type(YYLTYPE *loc, const glsl_type *base,
ast_array_specifier *array_specifier,
struct _mesa_glsl_parse_state *state)
{
const glsl_type *array_type = base;
if (array_specifier != NULL) {
if (base->is_array()) {
/* From page 19 (page 25) of the GLSL 1.20 spec:
*
* "Only one-dimensional arrays may be declared."
*/
if (!state->check_arrays_of_arrays_allowed(loc)) {
return glsl_type::error_type;
}
}
for (exec_node *node = array_specifier->array_dimensions.get_tail_raw();
!node->is_head_sentinel(); node = node->prev) {
unsigned array_size = process_array_size(node, state);
array_type = glsl_type::get_array_instance(array_type, array_size);
}
}
return array_type;
}
static bool
precision_qualifier_allowed(const glsl_type *type)
{
/* Precision qualifiers apply to floating point, integer and opaque
* types.
*
* Section 4.5.2 (Precision Qualifiers) of the GLSL 1.30 spec says:
* "Any floating point or any integer declaration can have the type
* preceded by one of these precision qualifiers [...] Literal
* constants do not have precision qualifiers. Neither do Boolean
* variables.
*
* Section 4.5 (Precision and Precision Qualifiers) of the GLSL 1.30
* spec also says:
*
* "Precision qualifiers are added for code portability with OpenGL
* ES, not for functionality. They have the same syntax as in OpenGL
* ES."
*
* Section 8 (Built-In Functions) of the GLSL ES 1.00 spec says:
*
* "uniform lowp sampler2D sampler;
* highp vec2 coord;
* ...
* lowp vec4 col = texture2D (sampler, coord);
* // texture2D returns lowp"
*
* From this, we infer that GLSL 1.30 (and later) should allow precision
* qualifiers on sampler types just like float and integer types.
*/
const glsl_type *const t = type->without_array();
return (t->is_float() || t->is_integer_32() || t->contains_opaque()) &&
!t->is_struct();
}
const glsl_type *
ast_type_specifier::glsl_type(const char **name,
struct _mesa_glsl_parse_state *state) const
{
const struct glsl_type *type;
if (this->type != NULL)
type = this->type;
else if (structure)
type = structure->type;
else
type = state->symbols->get_type(this->type_name);
*name = this->type_name;
YYLTYPE loc = this->get_location();
type = process_array_type(&loc, type, this->array_specifier, state);
return type;
}
/**
* From the OpenGL ES 3.0 spec, 4.5.4 Default Precision Qualifiers:
*
* "The precision statement
*
* precision precision-qualifier type;
*
* can be used to establish a default precision qualifier. The type field can
* be either int or float or any of the sampler types, (...) If type is float,
* the directive applies to non-precision-qualified floating point type
* (scalar, vector, and matrix) declarations. If type is int, the directive
* applies to all non-precision-qualified integer type (scalar, vector, signed,
* and unsigned) declarations."
*
* We use the symbol table to keep the values of the default precisions for
* each 'type' in each scope and we use the 'type' string from the precision
* statement as key in the symbol table. When we want to retrieve the default
* precision associated with a given glsl_type we need to know the type string
* associated with it. This is what this function returns.
*/
static const char *
get_type_name_for_precision_qualifier(const glsl_type *type)
{
switch (type->base_type) {
case GLSL_TYPE_FLOAT:
return "float";
case GLSL_TYPE_UINT:
case GLSL_TYPE_INT:
return "int";
case GLSL_TYPE_ATOMIC_UINT:
return "atomic_uint";
case GLSL_TYPE_IMAGE:
/* fallthrough */
case GLSL_TYPE_SAMPLER: {
const unsigned type_idx =
type->sampler_array + 2 * type->sampler_shadow;
const unsigned offset = type->is_sampler() ? 0 : 4;
assert(type_idx < 4);
switch (type->sampled_type) {
case GLSL_TYPE_FLOAT:
switch (type->sampler_dimensionality) {
case GLSL_SAMPLER_DIM_1D: {
assert(type->is_sampler());
static const char *const names[4] = {
"sampler1D", "sampler1DArray",
"sampler1DShadow", "sampler1DArrayShadow"
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_2D: {
static const char *const names[8] = {
"sampler2D", "sampler2DArray",
"sampler2DShadow", "sampler2DArrayShadow",
"image2D", "image2DArray", NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_3D: {
static const char *const names[8] = {
"sampler3D", NULL, NULL, NULL,
"image3D", NULL, NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_CUBE: {
static const char *const names[8] = {
"samplerCube", "samplerCubeArray",
"samplerCubeShadow", "samplerCubeArrayShadow",
"imageCube", NULL, NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_MS: {
assert(type->is_sampler());
static const char *const names[4] = {
"sampler2DMS", "sampler2DMSArray", NULL, NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_RECT: {
assert(type->is_sampler());
static const char *const names[4] = {
"samplerRect", NULL, "samplerRectShadow", NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_BUF: {
static const char *const names[8] = {
"samplerBuffer", NULL, NULL, NULL,
"imageBuffer", NULL, NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_EXTERNAL: {
assert(type->is_sampler());
static const char *const names[4] = {
"samplerExternalOES", NULL, NULL, NULL
};
return names[type_idx];
}
default:
unreachable("Unsupported sampler/image dimensionality");
} /* sampler/image float dimensionality */
break;
case GLSL_TYPE_INT:
switch (type->sampler_dimensionality) {
case GLSL_SAMPLER_DIM_1D: {
assert(type->is_sampler());
static const char *const names[4] = {
"isampler1D", "isampler1DArray", NULL, NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_2D: {
static const char *const names[8] = {
"isampler2D", "isampler2DArray", NULL, NULL,
"iimage2D", "iimage2DArray", NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_3D: {
static const char *const names[8] = {
"isampler3D", NULL, NULL, NULL,
"iimage3D", NULL, NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_CUBE: {
static const char *const names[8] = {
"isamplerCube", "isamplerCubeArray", NULL, NULL,
"iimageCube", NULL, NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_MS: {
assert(type->is_sampler());
static const char *const names[4] = {
"isampler2DMS", "isampler2DMSArray", NULL, NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_RECT: {
assert(type->is_sampler());
static const char *const names[4] = {
"isamplerRect", NULL, "isamplerRectShadow", NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_BUF: {
static const char *const names[8] = {
"isamplerBuffer", NULL, NULL, NULL,
"iimageBuffer", NULL, NULL, NULL
};
return names[offset + type_idx];
}
default:
unreachable("Unsupported isampler/iimage dimensionality");
} /* sampler/image int dimensionality */
break;
case GLSL_TYPE_UINT:
switch (type->sampler_dimensionality) {
case GLSL_SAMPLER_DIM_1D: {
assert(type->is_sampler());
static const char *const names[4] = {
"usampler1D", "usampler1DArray", NULL, NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_2D: {
static const char *const names[8] = {
"usampler2D", "usampler2DArray", NULL, NULL,
"uimage2D", "uimage2DArray", NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_3D: {
static const char *const names[8] = {
"usampler3D", NULL, NULL, NULL,
"uimage3D", NULL, NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_CUBE: {
static const char *const names[8] = {
"usamplerCube", "usamplerCubeArray", NULL, NULL,
"uimageCube", NULL, NULL, NULL
};
return names[offset + type_idx];
}
case GLSL_SAMPLER_DIM_MS: {
assert(type->is_sampler());
static const char *const names[4] = {
"usampler2DMS", "usampler2DMSArray", NULL, NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_RECT: {
assert(type->is_sampler());
static const char *const names[4] = {
"usamplerRect", NULL, "usamplerRectShadow", NULL
};
return names[type_idx];
}
case GLSL_SAMPLER_DIM_BUF: {
static const char *const names[8] = {
"usamplerBuffer", NULL, NULL, NULL,
"uimageBuffer", NULL, NULL, NULL
};
return names[offset + type_idx];
}
default:
unreachable("Unsupported usampler/uimage dimensionality");
} /* sampler/image uint dimensionality */
break;
default:
unreachable("Unsupported sampler/image type");
} /* sampler/image type */
break;
} /* GLSL_TYPE_SAMPLER/GLSL_TYPE_IMAGE */
break;
default:
unreachable("Unsupported type");
} /* base type */
}
static unsigned
select_gles_precision(unsigned qual_precision,
const glsl_type *type,
struct _mesa_glsl_parse_state *state, YYLTYPE *loc)
{
/* Precision qualifiers do not have any meaning in Desktop GLSL.
* In GLES we take the precision from the type qualifier if present,
* otherwise, if the type of the variable allows precision qualifiers at
* all, we look for the default precision qualifier for that type in the
* current scope.
*/
assert(state->es_shader);
unsigned precision = GLSL_PRECISION_NONE;
if (qual_precision) {
precision = qual_precision;
} else if (precision_qualifier_allowed(type)) {
const char *type_name =
get_type_name_for_precision_qualifier(type->without_array());
assert(type_name != NULL);
precision =
state->symbols->get_default_precision_qualifier(type_name);
if (precision == ast_precision_none) {
_mesa_glsl_error(loc, state,
"No precision specified in this scope for type `%s'",
type->name);
}
}
/* Section 4.1.7.3 (Atomic Counters) of the GLSL ES 3.10 spec says:
*
* "The default precision of all atomic types is highp. It is an error to
* declare an atomic type with a different precision or to specify the
* default precision for an atomic type to be lowp or mediump."
*/
if (type->is_atomic_uint() && precision != ast_precision_high) {
_mesa_glsl_error(loc, state,
"atomic_uint can only have highp precision qualifier");
}
return precision;
}
const glsl_type *
ast_fully_specified_type::glsl_type(const char **name,
struct _mesa_glsl_parse_state *state) const
{
return this->specifier->glsl_type(name, state);
}
/**
* Determine whether a toplevel variable declaration declares a varying. This
* function operates by examining the variable's mode and the shader target,
* so it correctly identifies linkage variables regardless of whether they are
* declared using the deprecated "varying" syntax or the new "in/out" syntax.
*
* Passing a non-toplevel variable declaration (e.g. a function parameter) to
* this function will produce undefined results.
*/
static bool
is_varying_var(ir_variable *var, gl_shader_stage target)
{
switch (target) {
case MESA_SHADER_VERTEX:
return var->data.mode == ir_var_shader_out;
case MESA_SHADER_FRAGMENT:
return var->data.mode == ir_var_shader_in ||
(var->data.mode == ir_var_system_value &&
var->data.location == SYSTEM_VALUE_FRAG_COORD);
default:
return var->data.mode == ir_var_shader_out || var->data.mode == ir_var_shader_in;
}
}
static bool
is_allowed_invariant(ir_variable *var, struct _mesa_glsl_parse_state *state)
{
if (is_varying_var(var, state->stage))
return true;
/* From Section 4.6.1 ("The Invariant Qualifier") GLSL 1.20 spec:
* "Only variables output from a vertex shader can be candidates
* for invariance".
*/
if (!state->is_version(130, 100))
return false;
/*
* Later specs remove this language - so allowed invariant
* on fragment shader outputs as well.
*/
if (state->stage == MESA_SHADER_FRAGMENT &&
var->data.mode == ir_var_shader_out)
return true;
return false;
}
/**
* Matrix layout qualifiers are only allowed on certain types
*/
static void
validate_matrix_layout_for_type(struct _mesa_glsl_parse_state *state,
YYLTYPE *loc,
const glsl_type *type,
ir_variable *var)
{
if (var && !var->is_in_buffer_block()) {
/* Layout qualifiers may only apply to interface blocks and fields in
* them.
*/
_mesa_glsl_error(loc, state,
"uniform block layout qualifiers row_major and "
"column_major may not be applied to variables "
"outside of uniform blocks");
} else if (!type->without_array()->is_matrix()) {
/* The OpenGL ES 3.0 conformance tests did not originally allow
* matrix layout qualifiers on non-matrices. However, the OpenGL
* 4.4 and OpenGL ES 3.0 (revision TBD) specifications were
* amended to specifically allow these layouts on all types. Emit
* a warning so that people know their code may not be portable.
*/
_mesa_glsl_warning(loc, state,
"uniform block layout qualifiers row_major and "
"column_major applied to non-matrix types may "
"be rejected by older compilers");
}
}
static bool
validate_xfb_buffer_qualifier(YYLTYPE *loc,
struct _mesa_glsl_parse_state *state,
unsigned xfb_buffer) {
if (xfb_buffer >= state->Const.MaxTransformFeedbackBuffers) {
_mesa_glsl_error(loc, state,
"invalid xfb_buffer specified %d is larger than "
"MAX_TRANSFORM_FEEDBACK_BUFFERS - 1 (%d).",
xfb_buffer,
state->Const.MaxTransformFeedbackBuffers - 1);
return false;
}
return true;
}
/* From the ARB_enhanced_layouts spec:
*
* "Variables and block members qualified with *xfb_offset* can be
* scalars, vectors, matrices, structures, and (sized) arrays of these.
* The offset must be a multiple of the size of the first component of
* the first qualified variable or block member, or a compile-time error
* results. Further, if applied to an aggregate containing a double,
* the offset must also be a multiple of 8, and the space taken in the
* buffer will be a multiple of 8.
*/
static bool
validate_xfb_offset_qualifier(YYLTYPE *loc,
struct _mesa_glsl_parse_state *state,
int xfb_offset, const glsl_type *type,
unsigned component_size) {
const glsl_type *t_without_array = type->without_array();
if (xfb_offset != -1 && type->is_unsized_array()) {
_mesa_glsl_error(loc, state,
"xfb_offset can't be used with unsized arrays.");
return false;
}
/* Make sure nested structs don't contain unsized arrays, and validate
* any xfb_offsets on interface members.
*/
if (t_without_array->is_struct() || t_without_array->is_interface())
for (unsigned int i = 0; i < t_without_array->length; i++) {
const glsl_type *member_t = t_without_array->fields.structure[i].type;
/* When the interface block doesn't have an xfb_offset qualifier then
* we apply the component size rules at the member level.
*/
if (xfb_offset == -1)
component_size = member_t->contains_double() ? 8 : 4;
int xfb_offset = t_without_array->fields.structure[i].offset;
validate_xfb_offset_qualifier(loc, state, xfb_offset, member_t,
component_size);
}
/* Nested structs or interface block without offset may not have had an
* offset applied yet so return.
*/
if (xfb_offset == -1) {
return true;
}
if (xfb_offset % component_size) {
_mesa_glsl_error(loc, state,
"invalid qualifier xfb_offset=%d must be a multiple "
"of the first component size of the first qualified "
"variable or block member. Or double if an aggregate "
"that contains a double (%d).",
xfb_offset, component_size);
return false;
}
return true;
}
static bool
validate_stream_qualifier(YYLTYPE *loc, struct _mesa_glsl_parse_state *state,
unsigned stream)
{
if (stream >= state->ctx->Const.MaxVertexStreams) {
_mesa_glsl_error(loc, state,
"invalid stream specified %d is larger than "
"MAX_VERTEX_STREAMS - 1 (%d).",
stream, state->ctx->Const.MaxVertexStreams - 1);
return false;
}
return true;
}
static void
apply_explicit_binding(struct _mesa_glsl_parse_state *state,
YYLTYPE *loc,
ir_variable *var,
const glsl_type *type,
const ast_type_qualifier *qual)
{
if (!qual->flags.q.uniform && !qual->flags.q.buffer) {
_mesa_glsl_error(loc, state,
"the \"binding\" qualifier only applies to uniforms and "
"shader storage buffer objects");
return;
}
unsigned qual_binding;
if (!process_qualifier_constant(state, loc, "binding", qual->binding,
&qual_binding)) {
return;
}
const struct gl_context *const ctx = state->ctx;
unsigned elements = type->is_array() ? type->arrays_of_arrays_size() : 1;
unsigned max_index = qual_binding + elements - 1;
const glsl_type *base_type = type->without_array();
if (base_type->is_interface()) {
/* UBOs. From page 60 of the GLSL 4.20 specification:
* "If the binding point for any uniform block instance is less than zero,
* or greater than or equal to the implementation-dependent maximum
* number of uniform buffer bindings, a compilation error will occur.
* When the binding identifier is used with a uniform block instanced as
* an array of size N, all elements of the array from binding through
* binding + N – 1 must be within this range."
*
* The implementation-dependent maximum is GL_MAX_UNIFORM_BUFFER_BINDINGS.
*/
if (qual->flags.q.uniform &&
max_index >= ctx->Const.MaxUniformBufferBindings) {
_mesa_glsl_error(loc, state, "layout(binding = %u) for %d UBOs exceeds "
"the maximum number of UBO binding points (%d)",
qual_binding, elements,
ctx->Const.MaxUniformBufferBindings);
return;
}
/* SSBOs. From page 67 of the GLSL 4.30 specification:
* "If the binding point for any uniform or shader storage block instance
* is less than zero, or greater than or equal to the
* implementation-dependent maximum number of uniform buffer bindings, a
* compile-time error will occur. When the binding identifier is used
* with a uniform or shader storage block instanced as an array of size
* N, all elements of the array from binding through binding + N – 1 must
* be within this range."
*/
if (qual->flags.q.buffer &&
max_index >= ctx->Const.MaxShaderStorageBufferBindings) {
_mesa_glsl_error(loc, state, "layout(binding = %u) for %d SSBOs exceeds "
"the maximum number of SSBO binding points (%d)",
qual_binding, elements,
ctx->Const.MaxShaderStorageBufferBindings);
return;
}
} else if (base_type->is_sampler()) {
/* Samplers. From page 63 of the GLSL 4.20 specification:
* "If the binding is less than zero, or greater than or equal to the
* implementation-dependent maximum supported number of units, a
* compilation error will occur. When the binding identifier is used
* with an array of size N, all elements of the array from binding
* through binding + N - 1 must be within this range."
*/
unsigned limit = ctx->Const.MaxCombinedTextureImageUnits;
if (max_index >= limit) {
_mesa_glsl_error(loc, state, "layout(binding = %d) for %d samplers "
"exceeds the maximum number of texture image units "
"(%u)", qual_binding, elements, limit);
return;
}
} else if (base_type->contains_atomic()) {
assert(ctx->Const.MaxAtomicBufferBindings <= MAX_COMBINED_ATOMIC_BUFFERS);
if (qual_binding >= ctx->Const.MaxAtomicBufferBindings) {
_mesa_glsl_error(loc, state, "layout(binding = %d) exceeds the "
"maximum number of atomic counter buffer bindings "
"(%u)", qual_binding,
ctx->Const.MaxAtomicBufferBindings);
return;
}
} else if ((state->is_version(420, 310) ||
state->ARB_shading_language_420pack_enable) &&
base_type->is_image()) {
assert(ctx->Const.MaxImageUnits <= MAX_IMAGE_UNITS);
if (max_index >= ctx->Const.MaxImageUnits) {
_mesa_glsl_error(loc, state, "Image binding %d exceeds the "
"maximum number of image units (%d)", max_index,
ctx->Const.MaxImageUnits);
return;
}
} else {
_mesa_glsl_error(loc, state,
"the \"binding\" qualifier only applies to uniform "
"blocks, storage blocks, opaque variables, or arrays "
"thereof");
return;
}
var->data.explicit_binding = true;
var->data.binding = qual_binding;
return;
}
static void
validate_fragment_flat_interpolation_input(struct _mesa_glsl_parse_state *state,
YYLTYPE *loc,
const glsl_interp_mode interpolation,
const struct glsl_type *var_type,
ir_variable_mode mode)
{
if (state->stage != MESA_SHADER_FRAGMENT ||
interpolation == INTERP_MODE_FLAT ||
mode != ir_var_shader_in)
return;
/* Integer fragment inputs must be qualified with 'flat'. In GLSL ES,
* so must integer vertex outputs.
*
* From section 4.3.4 ("Inputs") of the GLSL 1.50 spec:
* "Fragment shader inputs that are signed or unsigned integers or
* integer vectors must be qualified with the interpolation qualifier
* flat."
*
* From section 4.3.4 ("Input Variables") of the GLSL 3.00 ES spec:
* "Fragment shader inputs that are, or contain, signed or unsigned
* integers or integer vectors must be qualified with the
* interpolation qualifier flat."
*
* From section 4.3.6 ("Output Variables") of the GLSL 3.00 ES spec:
* "Vertex shader outputs that are, or contain, signed or unsigned
* integers or integer vectors must be qualified with the
* interpolation qualifier flat."
*
* Note that prior to GLSL 1.50, this requirement applied to vertex
* outputs rather than fragment inputs. That creates problems in the
* presence of geometry shaders, so we adopt the GLSL 1.50 rule for all
* desktop GL shaders. For GLSL ES shaders, we follow the spec and
* apply the restriction to both vertex outputs and fragment inputs.
*
* Note also that the desktop GLSL specs are missing the text "or
* contain"; this is presumably an oversight, since there is no
* reasonable way to interpolate a fragment shader input that contains
* an integer. See Khronos bug #15671.
*/
if ((state->is_version(130, 300) || state->EXT_gpu_shader4_enable)
&& var_type->contains_integer()) {
_mesa_glsl_error(loc, state, "if a fragment input is (or contains) "
"an integer, then it must be qualified with 'flat'");
}
/* Double fragment inputs must be qualified with 'flat'.
*
* From the "Overview" of the ARB_gpu_shader_fp64 extension spec:
* "This extension does not support interpolation of double-precision
* values; doubles used as fragment shader inputs must be qualified as
* "flat"."
*
* From section 4.3.4 ("Inputs") of the GLSL 4.00 spec:
* "Fragment shader inputs that are signed or unsigned integers, integer
* vectors, or any double-precision floating-point type must be
* qualified with the interpolation qualifier flat."
*
* Note that the GLSL specs are missing the text "or contain"; this is
* presumably an oversight. See Khronos bug #15671.
*
* The 'double' type does not exist in GLSL ES so far.
*/
if (state->has_double()
&& var_type->contains_double()) {
_mesa_glsl_error(loc, state, "if a fragment input is (or contains) "
"a double, then it must be qualified with 'flat'");
}
/* Bindless sampler/image fragment inputs must be qualified with 'flat'.
*
* From section 4.3.4 of the ARB_bindless_texture spec:
*
* "(modify last paragraph, p. 35, allowing samplers and images as
* fragment shader inputs) ... Fragment inputs can only be signed and
* unsigned integers and integer vectors, floating point scalars,
* floating-point vectors, matrices, sampler and image types, or arrays
* or structures of these. Fragment shader inputs that are signed or
* unsigned integers, integer vectors, or any double-precision floating-
* point type, or any sampler or image type must be qualified with the
* interpolation qualifier "flat"."
*/
if (state->has_bindless()
&& (var_type->contains_sampler() || var_type->contains_image())) {
_mesa_glsl_error(loc, state, "if a fragment input is (or contains) "
"a bindless sampler (or image), then it must be "
"qualified with 'flat'");
}
}
static void
validate_interpolation_qualifier(struct _mesa_glsl_parse_state *state,
YYLTYPE *loc,
const glsl_interp_mode interpolation,
const struct ast_type_qualifier *qual,
const struct glsl_type *var_type,
ir_variable_mode mode)
{
/* Interpolation qualifiers can only apply to shader inputs or outputs, but
* not to vertex shader inputs nor fragment shader outputs.
*
* From section 4.3 ("Storage Qualifiers") of the GLSL 1.30 spec:
* "Outputs from a vertex shader (out) and inputs to a fragment
* shader (in) can be further qualified with one or more of these
* interpolation qualifiers"
* ...
* "These interpolation qualifiers may only precede the qualifiers in,
* centroid in, out, or centroid out in a declaration. They do not apply
* to the deprecated storage qualifiers varying or centroid
* varying. They also do not apply to inputs into a vertex shader or
* outputs from a fragment shader."
*
* From section 4.3 ("Storage Qualifiers") of the GLSL ES 3.00 spec:
* "Outputs from a shader (out) and inputs to a shader (in) can be
* further qualified with one of these interpolation qualifiers."
* ...
* "These interpolation qualifiers may only precede the qualifiers
* in, centroid in, out, or centroid out in a declaration. They do
* not apply to inputs into a vertex shader or outputs from a
* fragment shader."
*/
if ((state->is_version(130, 300) || state->EXT_gpu_shader4_enable)
&& interpolation != INTERP_MODE_NONE) {
const char *i = interpolation_string(interpolation);
if (mode != ir_var_shader_in && mode != ir_var_shader_out)
_mesa_glsl_error(loc, state,
"interpolation qualifier `%s' can only be applied to "
"shader inputs or outputs.", i);
switch (state->stage) {
case MESA_SHADER_VERTEX:
if (mode == ir_var_shader_in) {
_mesa_glsl_error(loc, state,
"interpolation qualifier '%s' cannot be applied to "
"vertex shader inputs", i);
}
break;
case MESA_SHADER_FRAGMENT:
if (mode == ir_var_shader_out) {
_mesa_glsl_error(loc, state,
"interpolation qualifier '%s' cannot be applied to "
"fragment shader outputs", i);
}
break;
default:
break;
}
}
/* Interpolation qualifiers cannot be applied to 'centroid' and
* 'centroid varying'.
*
* From section 4.3 ("Storage Qualifiers") of the GLSL 1.30 spec:
* "interpolation qualifiers may only precede the qualifiers in,
* centroid in, out, or centroid out in a declaration. They do not apply
* to the deprecated storage qualifiers varying or centroid varying."
*
* These deprecated storage qualifiers do not exist in GLSL ES 3.00.
*
* GL_EXT_gpu_shader4 allows this.
*/
if (state->is_version(130, 0) && !state->EXT_gpu_shader4_enable
&& interpolation != INTERP_MODE_NONE
&& qual->flags.q.varying) {
const char *i = interpolation_string(interpolation);
const char *s;
if (qual->flags.q.centroid)
s = "centroid varying";
else
s = "varying";
_mesa_glsl_error(loc, state,
"qualifier '%s' cannot be applied to the "
"deprecated storage qualifier '%s'", i, s);
}
validate_fragment_flat_interpolation_input(state, loc, interpolation,
var_type, mode);
}
static glsl_interp_mode
interpret_interpolation_qualifier(const struct ast_type_qualifier *qual,
const struct glsl_type *var_type,
ir_variable_mode mode,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc)
{
glsl_interp_mode interpolation;
if (qual->flags.q.flat)
interpolation = INTERP_MODE_FLAT;
else if (qual->flags.q.noperspective)
interpolation = INTERP_MODE_NOPERSPECTIVE;
else if (qual->flags.q.smooth)
interpolation = INTERP_MODE_SMOOTH;
else
interpolation = INTERP_MODE_NONE;
validate_interpolation_qualifier(state, loc,
interpolation,
qual, var_type, mode);
return interpolation;
}
static void
apply_explicit_location(const struct ast_type_qualifier *qual,
ir_variable *var,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc)
{
bool fail = false;
unsigned qual_location;
if (!process_qualifier_constant(state, loc, "location", qual->location,
&qual_location)) {
return;
}
/* Checks for GL_ARB_explicit_uniform_location. */
if (qual->flags.q.uniform) {
if (!state->check_explicit_uniform_location_allowed(loc, var))
return;
const struct gl_context *const ctx = state->ctx;
unsigned max_loc = qual_location + var->type->uniform_locations() - 1;
if (max_loc >= ctx->Const.MaxUserAssignableUniformLocations) {
_mesa_glsl_error(loc, state, "location(s) consumed by uniform %s "
">= MAX_UNIFORM_LOCATIONS (%u)", var->name,
ctx->Const.MaxUserAssignableUniformLocations);
return;
}
var->data.explicit_location = true;
var->data.location = qual_location;
return;
}
/* Between GL_ARB_explicit_attrib_location an
* GL_ARB_separate_shader_objects, the inputs and outputs of any shader
* stage can be assigned explicit locations. The checking here associates
* the correct extension with the correct stage's input / output:
*
* input output
* ----- ------
* vertex explicit_loc sso
* tess control sso sso
* tess eval sso sso
* geometry sso sso
* fragment sso explicit_loc
*/
switch (state->stage) {
case MESA_SHADER_VERTEX:
if (var->data.mode == ir_var_shader_in) {
if (!state->check_explicit_attrib_location_allowed(loc, var))
return;
break;
}
if (var->data.mode == ir_var_shader_out) {
if (!state->check_separate_shader_objects_allowed(loc, var))
return;
break;
}
fail = true;
break;
case MESA_SHADER_TESS_CTRL:
case MESA_SHADER_TESS_EVAL:
case MESA_SHADER_GEOMETRY:
if (var->data.mode == ir_var_shader_in || var->data.mode == ir_var_shader_out) {
if (!state->check_separate_shader_objects_allowed(loc, var))
return;
break;
}
fail = true;
break;
case MESA_SHADER_FRAGMENT:
if (var->data.mode == ir_var_shader_in) {
if (!state->check_separate_shader_objects_allowed(loc, var))
return;
break;
}
if (var->data.mode == ir_var_shader_out) {
if (!state->check_explicit_attrib_location_allowed(loc, var))
return;
break;
}
fail = true;
break;
case MESA_SHADER_COMPUTE:
_mesa_glsl_error(loc, state,
"compute shader variables cannot be given "
"explicit locations");
return;
default:
fail = true;
break;
};
if (fail) {
_mesa_glsl_error(loc, state,
"%s cannot be given an explicit location in %s shader",
mode_string(var),
_mesa_shader_stage_to_string(state->stage));
} else {
var->data.explicit_location = true;
switch (state->stage) {
case MESA_SHADER_VERTEX:
var->data.location = (var->data.mode == ir_var_shader_in)
? (qual_location + VERT_ATTRIB_GENERIC0)
: (qual_location + VARYING_SLOT_VAR0);
break;
case MESA_SHADER_TESS_CTRL:
case MESA_SHADER_TESS_EVAL:
case MESA_SHADER_GEOMETRY:
if (var->data.patch)
var->data.location = qual_location + VARYING_SLOT_PATCH0;
else
var->data.location = qual_location + VARYING_SLOT_VAR0;
break;
case MESA_SHADER_FRAGMENT:
var->data.location = (var->data.mode == ir_var_shader_out)
? (qual_location + FRAG_RESULT_DATA0)
: (qual_location + VARYING_SLOT_VAR0);
break;
default:
assert(!"Unexpected shader type");
break;
}
/* Check if index was set for the uniform instead of the function */
if (qual->flags.q.explicit_index && qual->is_subroutine_decl()) {
_mesa_glsl_error(loc, state, "an index qualifier can only be "
"used with subroutine functions");
return;
}
unsigned qual_index;
if (qual->flags.q.explicit_index &&
process_qualifier_constant(state, loc, "index", qual->index,
&qual_index)) {
/* From the GLSL 4.30 specification, section 4.4.2 (Output
* Layout Qualifiers):
*
* "It is also a compile-time error if a fragment shader
* sets a layout index to less than 0 or greater than 1."
*
* Older specifications don't mandate a behavior; we take
* this as a clarification and always generate the error.
*/
if (qual_index > 1) {
_mesa_glsl_error(loc, state,
"explicit index may only be 0 or 1");
} else {
var->data.explicit_index = true;
var->data.index = qual_index;
}
}
}
}
static bool
validate_storage_for_sampler_image_types(ir_variable *var,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc)
{
/* From section 4.1.7 of the GLSL 4.40 spec:
*
* "[Opaque types] can only be declared as function
* parameters or uniform-qualified variables."
*
* From section 4.1.7 of the ARB_bindless_texture spec:
*
* "Samplers may be declared as shader inputs and outputs, as uniform
* variables, as temporary variables, and as function parameters."
*
* From section 4.1.X of the ARB_bindless_texture spec:
*
* "Images may be declared as shader inputs and outputs, as uniform
* variables, as temporary variables, and as function parameters."
*/
if (state->has_bindless()) {
if (var->data.mode != ir_var_auto &&
var->data.mode != ir_var_uniform &&
var->data.mode != ir_var_shader_in &&
var->data.mode != ir_var_shader_out &&
var->data.mode != ir_var_function_in &&
var->data.mode != ir_var_function_out &&
var->data.mode != ir_var_function_inout) {
_mesa_glsl_error(loc, state, "bindless image/sampler variables may "
"only be declared as shader inputs and outputs, as "
"uniform variables, as temporary variables and as "
"function parameters");
return false;
}
} else {
if (var->data.mode != ir_var_uniform &&
var->data.mode != ir_var_function_in) {
_mesa_glsl_error(loc, state, "image/sampler variables may only be "
"declared as function parameters or "
"uniform-qualified global variables");
return false;
}
}
return true;
}
static bool
validate_memory_qualifier_for_type(struct _mesa_glsl_parse_state *state,
YYLTYPE *loc,
const struct ast_type_qualifier *qual,
const glsl_type *type)
{
/* From Section 4.10 (Memory Qualifiers) of the GLSL 4.50 spec:
*
* "Memory qualifiers are only supported in the declarations of image
* variables, buffer variables, and shader storage blocks; it is an error
* to use such qualifiers in any other declarations.
*/
if (!type->is_image() && !qual->flags.q.buffer) {
if (qual->flags.q.read_only ||
qual->flags.q.write_only ||
qual->flags.q.coherent ||
qual->flags.q._volatile ||
qual->flags.q.restrict_flag) {
_mesa_glsl_error(loc, state, "memory qualifiers may only be applied "
"in the declarations of image variables, buffer "
"variables, and shader storage blocks");
return false;
}
}
return true;
}
static bool
validate_image_format_qualifier_for_type(struct _mesa_glsl_parse_state *state,
YYLTYPE *loc,
const struct ast_type_qualifier *qual,
const glsl_type *type)
{
/* From section 4.4.6.2 (Format Layout Qualifiers) of the GLSL 4.50 spec:
*
* "Format layout qualifiers can be used on image variable declarations
* (those declared with a basic type having “image ” in its keyword)."
*/
if (!type->is_image() && qual->flags.q.explicit_image_format) {
_mesa_glsl_error(loc, state, "format layout qualifiers may only be "
"applied to images");
return false;
}
return true;
}
static void
apply_image_qualifier_to_variable(const struct ast_type_qualifier *qual,
ir_variable *var,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc)
{
const glsl_type *base_type = var->type->without_array();
if (!validate_image_format_qualifier_for_type(state, loc, qual, base_type) ||
!validate_memory_qualifier_for_type(state, loc, qual, base_type))
return;
if (!base_type->is_image())
return;
if (!validate_storage_for_sampler_image_types(var, state, loc))
return;
var->data.memory_read_only |= qual->flags.q.read_only;
var->data.memory_write_only |= qual->flags.q.write_only;
var->data.memory_coherent |= qual->flags.q.coherent;
var->data.memory_volatile |= qual->flags.q._volatile;
var->data.memory_restrict |= qual->flags.q.restrict_flag;
if (qual->flags.q.explicit_image_format) {
if (var->data.mode == ir_var_function_in) {
_mesa_glsl_error(loc, state, "format qualifiers cannot be used on "
"image function parameters");
}
if (qual->image_base_type != base_type->sampled_type) {
_mesa_glsl_error(loc, state, "format qualifier doesn't match the base "
"data type of the image");
}
var->data.image_format = qual->image_format;
} else if (state->has_image_load_formatted()) {
if (var->data.mode == ir_var_uniform &&
state->EXT_shader_image_load_formatted_warn) {
_mesa_glsl_warning(loc, state, "GL_EXT_image_load_formatted used");
}
} else {
if (var->data.mode == ir_var_uniform) {
if (state->es_shader ||
!(state->is_version(420, 310) || state->ARB_shader_image_load_store_enable)) {
_mesa_glsl_error(loc, state, "all image uniforms must have a "
"format layout qualifier");
} else if (!qual->flags.q.write_only) {
_mesa_glsl_error(loc, state, "image uniforms not qualified with "
"`writeonly' must have a format layout qualifier");
}
}
var->data.image_format = PIPE_FORMAT_NONE;
}
/* From page 70 of the GLSL ES 3.1 specification:
*
* "Except for image variables qualified with the format qualifiers r32f,
* r32i, and r32ui, image variables must specify either memory qualifier
* readonly or the memory qualifier writeonly."
*/
if (state->es_shader &&
var->data.image_format != PIPE_FORMAT_R32_FLOAT &&
var->data.image_format != PIPE_FORMAT_R32_SINT &&
var->data.image_format != PIPE_FORMAT_R32_UINT &&
!var->data.memory_read_only &&
!var->data.memory_write_only) {
_mesa_glsl_error(loc, state, "image variables of format other than r32f, "
"r32i or r32ui must be qualified `readonly' or "
"`writeonly'");
}
}
static inline const char*
get_layout_qualifier_string(bool origin_upper_left, bool pixel_center_integer)
{
if (origin_upper_left && pixel_center_integer)
return "origin_upper_left, pixel_center_integer";
else if (origin_upper_left)
return "origin_upper_left";
else if (pixel_center_integer)
return "pixel_center_integer";
else
return " ";
}
static inline bool
is_conflicting_fragcoord_redeclaration(struct _mesa_glsl_parse_state *state,
const struct ast_type_qualifier *qual)
{
/* If gl_FragCoord was previously declared, and the qualifiers were
* different in any way, return true.
*/
if (state->fs_redeclares_gl_fragcoord) {
return (state->fs_pixel_center_integer != qual->flags.q.pixel_center_integer
|| state->fs_origin_upper_left != qual->flags.q.origin_upper_left);
}
return false;
}
static inline bool
is_conflicting_layer_redeclaration(struct _mesa_glsl_parse_state *state,
const struct ast_type_qualifier *qual)
{
if (state->redeclares_gl_layer) {
return state->layer_viewport_relative != qual->flags.q.viewport_relative;
}
return false;
}
static inline void
validate_array_dimensions(const glsl_type *t,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc) {
if (t->is_array()) {
t = t->fields.array;
while (t->is_array()) {
if (t->is_unsized_array()) {
_mesa_glsl_error(loc, state,
"only the outermost array dimension can "
"be unsized",
t->name);
break;
}
t = t->fields.array;
}
}
}
static void
apply_bindless_qualifier_to_variable(const struct ast_type_qualifier *qual,
ir_variable *var,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc)
{
bool has_local_qualifiers = qual->flags.q.bindless_sampler ||
qual->flags.q.bindless_image ||
qual->flags.q.bound_sampler ||
qual->flags.q.bound_image;
/* The ARB_bindless_texture spec says:
*
* "Modify Section 4.4.6 Opaque-Uniform Layout Qualifiers of the GLSL 4.30
* spec"
*
* "If these layout qualifiers are applied to other types of default block
* uniforms, or variables with non-uniform storage, a compile-time error
* will be generated."
*/
if (has_local_qualifiers && !qual->flags.q.uniform) {
_mesa_glsl_error(loc, state, "ARB_bindless_texture layout qualifiers "
"can only be applied to default block uniforms or "
"variables with uniform storage");
return;
}
/* The ARB_bindless_texture spec doesn't state anything in this situation,
* but it makes sense to only allow bindless_sampler/bound_sampler for
* sampler types, and respectively bindless_image/bound_image for image
* types.
*/
if ((qual->flags.q.bindless_sampler || qual->flags.q.bound_sampler) &&
!var->type->contains_sampler()) {
_mesa_glsl_error(loc, state, "bindless_sampler or bound_sampler can only "
"be applied to sampler types");
return;
}
if ((qual->flags.q.bindless_image || qual->flags.q.bound_image) &&
!var->type->contains_image()) {
_mesa_glsl_error(loc, state, "bindless_image or bound_image can only be "
"applied to image types");
return;
}
/* The bindless_sampler/bindless_image (and respectively
* bound_sampler/bound_image) layout qualifiers can be set at global and at
* local scope.
*/
if (var->type->contains_sampler() || var->type->contains_image()) {
var->data.bindless = qual->flags.q.bindless_sampler ||
qual->flags.q.bindless_image ||
state->bindless_sampler_specified ||
state->bindless_image_specified;
var->data.bound = qual->flags.q.bound_sampler ||
qual->flags.q.bound_image ||
state->bound_sampler_specified ||
state->bound_image_specified;
}
}
static void
apply_layout_qualifier_to_variable(const struct ast_type_qualifier *qual,
ir_variable *var,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc)
{
if (var->name != NULL && strcmp(var->name, "gl_FragCoord") == 0) {
/* Section 4.3.8.1, page 39 of GLSL 1.50 spec says:
*
* "Within any shader, the first redeclarations of gl_FragCoord
* must appear before any use of gl_FragCoord."
*
* Generate a compiler error if above condition is not met by the
* fragment shader.
*/
ir_variable *earlier = state->symbols->get_variable("gl_FragCoord");
if (earlier != NULL &&
earlier->data.used &&
!state->fs_redeclares_gl_fragcoord) {
_mesa_glsl_error(loc, state,
"gl_FragCoord used before its first redeclaration "
"in fragment shader");
}
/* Make sure all gl_FragCoord redeclarations specify the same layout
* qualifiers.
*/
if (is_conflicting_fragcoord_redeclaration(state, qual)) {
const char *const qual_string =
get_layout_qualifier_string(qual->flags.q.origin_upper_left,
qual->flags.q.pixel_center_integer);
const char *const state_string =
get_layout_qualifier_string(state->fs_origin_upper_left,
state->fs_pixel_center_integer);
_mesa_glsl_error(loc, state,
"gl_FragCoord redeclared with different layout "
"qualifiers (%s) and (%s) ",
state_string,
qual_string);
}
state->fs_origin_upper_left = qual->flags.q.origin_upper_left;
state->fs_pixel_center_integer = qual->flags.q.pixel_center_integer;
state->fs_redeclares_gl_fragcoord_with_no_layout_qualifiers =
!qual->flags.q.origin_upper_left && !qual->flags.q.pixel_center_integer;
state->fs_redeclares_gl_fragcoord =
state->fs_origin_upper_left ||
state->fs_pixel_center_integer ||
state->fs_redeclares_gl_fragcoord_with_no_layout_qualifiers;
}
if ((qual->flags.q.origin_upper_left || qual->flags.q.pixel_center_integer)
&& (strcmp(var->name, "gl_FragCoord") != 0)) {
const char *const qual_string = (qual->flags.q.origin_upper_left)
? "origin_upper_left" : "pixel_center_integer";
_mesa_glsl_error(loc, state,
"layout qualifier `%s' can only be applied to "
"fragment shader input `gl_FragCoord'",
qual_string);
}
if (qual->flags.q.explicit_location) {
apply_explicit_location(qual, var, state, loc);
if (qual->flags.q.explicit_component) {
unsigned qual_component;
if (process_qualifier_constant(state, loc, "component",
qual->component, &qual_component)) {
const glsl_type *type = var->type->without_array();
unsigned components = type->component_slots();
if (type->is_matrix() || type->is_struct()) {
_mesa_glsl_error(loc, state, "component layout qualifier "
"cannot be applied to a matrix, a structure, "
"a block, or an array containing any of "
"these.");
} else if (components > 4 && type->is_64bit()) {
_mesa_glsl_error(loc, state, "component layout qualifier "
"cannot be applied to dvec%u.",
components / 2);
} else if (qual_component != 0 &&
(qual_component + components - 1) > 3) {
_mesa_glsl_error(loc, state, "component overflow (%u > 3)",
(qual_component + components - 1));
} else if (qual_component == 1 && type->is_64bit()) {
/* We don't bother checking for 3 as it should be caught by the
* overflow check above.
*/
_mesa_glsl_error(loc, state, "doubles cannot begin at "
"component 1 or 3");
} else {
var->data.explicit_component = true;
var->data.location_frac = qual_component;
}
}
}
} else if (qual->flags.q.explicit_index) {
if (!qual->subroutine_list)
_mesa_glsl_error(loc, state,
"explicit index requires explicit location");
} else if (qual->flags.q.explicit_component) {
_mesa_glsl_error(loc, state,
"explicit component requires explicit location");
}
if (qual->flags.q.explicit_binding) {
apply_explicit_binding(state, loc, var, var->type, qual);
}
if (state->stage == MESA_SHADER_GEOMETRY &&
qual->flags.q.out && qual->flags.q.stream) {
unsigned qual_stream;
if (process_qualifier_constant(state, loc, "stream", qual->stream,
&qual_stream) &&
validate_stream_qualifier(loc, state, qual_stream)) {
var->data.stream = qual_stream;
}
}
if (qual->flags.q.out && qual->flags.q.xfb_buffer) {
unsigned qual_xfb_buffer;
if (process_qualifier_constant(state, loc, "xfb_buffer",
qual->xfb_buffer, &qual_xfb_buffer) &&
validate_xfb_buffer_qualifier(loc, state, qual_xfb_buffer)) {
var->data.xfb_buffer = qual_xfb_buffer;
if (qual->flags.q.explicit_xfb_buffer)
var->data.explicit_xfb_buffer = true;
}
}
if (qual->flags.q.explicit_xfb_offset) {
unsigned qual_xfb_offset;
unsigned component_size = var->type->contains_double() ? 8 : 4;
if (process_qualifier_constant(state, loc, "xfb_offset",
qual->offset, &qual_xfb_offset) &&
validate_xfb_offset_qualifier(loc, state, (int) qual_xfb_offset,
var->type, component_size)) {
var->data.offset = qual_xfb_offset;
var->data.explicit_xfb_offset = true;
}
}
if (qual->flags.q.explicit_xfb_stride) {
unsigned qual_xfb_stride;
if (process_qualifier_constant(state, loc, "xfb_stride",
qual->xfb_stride, &qual_xfb_stride)) {
var->data.xfb_stride = qual_xfb_stride;
var->data.explicit_xfb_stride = true;
}
}
if (var->type->contains_atomic()) {
if (var->data.mode == ir_var_uniform) {
if (var->data.explicit_binding) {
unsigned *offset =
&state->atomic_counter_offsets[var->data.binding];
if (*offset % ATOMIC_COUNTER_SIZE)
_mesa_glsl_error(loc, state,
"misaligned atomic counter offset");
var->data.offset = *offset;
*offset += var->type->atomic_size();
} else {
_mesa_glsl_error(loc, state,
"atomic counters require explicit binding point");
}
} else if (var->data.mode != ir_var_function_in) {
_mesa_glsl_error(loc, state, "atomic counters may only be declared as "
"function parameters or uniform-qualified "
"global variables");
}
}
if (var->type->contains_sampler() &&
!validate_storage_for_sampler_image_types(var, state, loc))
return;
/* Is the 'layout' keyword used with parameters that allow relaxed checking.
* Many implementations of GL_ARB_fragment_coord_conventions_enable and some
* implementations (only Mesa?) GL_ARB_explicit_attrib_location_enable
* allowed the layout qualifier to be used with 'varying' and 'attribute'.
* These extensions and all following extensions that add the 'layout'
* keyword have been modified to require the use of 'in' or 'out'.
*
* The following extension do not allow the deprecated keywords:
*
* GL_AMD_conservative_depth
* GL_ARB_conservative_depth
* GL_ARB_gpu_shader5
* GL_ARB_separate_shader_objects
* GL_ARB_tessellation_shader
* GL_ARB_transform_feedback3
* GL_ARB_uniform_buffer_object
*
* It is unknown whether GL_EXT_shader_image_load_store or GL_NV_gpu_shader5
* allow layout with the deprecated keywords.
*/
const bool relaxed_layout_qualifier_checking =
state->ARB_fragment_coord_conventions_enable;
const bool uses_deprecated_qualifier = qual->flags.q.attribute
|| qual->flags.q.varying;
if (qual->has_layout() && uses_deprecated_qualifier) {
if (relaxed_layout_qualifier_checking) {
_mesa_glsl_warning(loc, state,
"`layout' qualifier may not be used with "
"`attribute' or `varying'");
} else {
_mesa_glsl_error(loc, state,
"`layout' qualifier may not be used with "
"`attribute' or `varying'");
}
}
/* Layout qualifiers for gl_FragDepth, which are enabled by extension
* AMD_conservative_depth.
*/
if (qual->flags.q.depth_type
&& !state->is_version(420, 0)
&& !state->AMD_conservative_depth_enable
&& !state->ARB_conservative_depth_enable) {
_mesa_glsl_error(loc, state,
"extension GL_AMD_conservative_depth or "
"GL_ARB_conservative_depth must be enabled "
"to use depth layout qualifiers");
} else if (qual->flags.q.depth_type
&& strcmp(var->name, "gl_FragDepth") != 0) {
_mesa_glsl_error(loc, state,
"depth layout qualifiers can be applied only to "
"gl_FragDepth");
}
switch (qual->depth_type) {
case ast_depth_any:
var->data.depth_layout = ir_depth_layout_any;
break;
case ast_depth_greater:
var->data.depth_layout = ir_depth_layout_greater;
break;
case ast_depth_less:
var->data.depth_layout = ir_depth_layout_less;
break;
case ast_depth_unchanged:
var->data.depth_layout = ir_depth_layout_unchanged;
break;
default:
var->data.depth_layout = ir_depth_layout_none;
break;
}
if (qual->flags.q.std140 ||
qual->flags.q.std430 ||
qual->flags.q.packed ||
qual->flags.q.shared) {
_mesa_glsl_error(loc, state,
"uniform and shader storage block layout qualifiers "
"std140, std430, packed, and shared can only be "
"applied to uniform or shader storage blocks, not "
"members");
}
if (qual->flags.q.row_major || qual->flags.q.column_major) {
validate_matrix_layout_for_type(state, loc, var->type, var);
}
/* From section 4.4.1.3 of the GLSL 4.50 specification (Fragment Shader
* Inputs):
*
* "Fragment shaders also allow the following layout qualifier on in only
* (not with variable declarations)
* layout-qualifier-id
* early_fragment_tests
* [...]"
*/
if (qual->flags.q.early_fragment_tests) {
_mesa_glsl_error(loc, state, "early_fragment_tests layout qualifier only "
"valid in fragment shader input layout declaration.");
}
if (qual->flags.q.inner_coverage) {
_mesa_glsl_error(loc, state, "inner_coverage layout qualifier only "
"valid in fragment shader input layout declaration.");
}
if (qual->flags.q.post_depth_coverage) {
_mesa_glsl_error(loc, state, "post_depth_coverage layout qualifier only "
"valid in fragment shader input layout declaration.");
}
if (state->has_bindless())
apply_bindless_qualifier_to_variable(qual, var, state, loc);
if (qual->flags.q.pixel_interlock_ordered ||
qual->flags.q.pixel_interlock_unordered ||
qual->flags.q.sample_interlock_ordered ||
qual->flags.q.sample_interlock_unordered) {
_mesa_glsl_error(loc, state, "interlock layout qualifiers: "
"pixel_interlock_ordered, pixel_interlock_unordered, "
"sample_interlock_ordered and sample_interlock_unordered, "
"only valid in fragment shader input layout declaration.");
}
if (var->name != NULL && strcmp(var->name, "gl_Layer") == 0) {
if (is_conflicting_layer_redeclaration(state, qual)) {
_mesa_glsl_error(loc, state, "gl_Layer redeclaration with "
"different viewport_relative setting than earlier");
}
state->redeclares_gl_layer = 1;
if (qual->flags.q.viewport_relative) {
state->layer_viewport_relative = 1;
}
} else if (qual->flags.q.viewport_relative) {
_mesa_glsl_error(loc, state,
"viewport_relative qualifier "
"can only be applied to gl_Layer.");
}
}
static void
apply_type_qualifier_to_variable(const struct ast_type_qualifier *qual,
ir_variable *var,
struct _mesa_glsl_parse_state *state,
YYLTYPE *loc,
bool is_parameter)
{
STATIC_ASSERT(sizeof(qual->flags.q) <= sizeof(qual->flags.i));
if (qual->flags.q.invariant) {
if (var->data.used) {
_mesa_glsl_error(loc, state,
"variable `%s' may not be redeclared "
"`invariant' after being used",
var->name);
} else {
var->data.explicit_invariant = true;
var->data.invariant = true;
}
}
if (qual->flags.q.precise) {
if (var->data.used) {
_mesa_glsl_error(loc, state,
"variable `%s' may not be redeclared "
"`precise' after being used",
var->name);
} else {
var->data.precise = 1;
}
}
if (qual->is_subroutine_decl() && !qual->flags.q.uniform) {
_mesa_glsl_error(loc, state,
"`subroutine' may only be applied to uniforms, "
"subroutine type declarations, or function definitions");
}
if (qual->flags.q.constant || qual->flags.q.attribute
|| qual->flags.q.uniform
|| (qual->flags.q.varying && (state->stage == MESA_SHADER_FRAGMENT)))
var->data.read_only = 1;
if (qual->flags.q.centroid)
var->data.centroid = 1;
if (qual->flags.q.sample)
var->data.sample = 1;
/* Precision qualifiers do not hold any meaning in Desktop GLSL */
if (state->es_shader) {
var->data.precision =
select_gles_precision(qual->precision, var->type, state, loc);
}
if (qual->flags.q.patch)
var->data.patch = 1;
if (qual->flags.q.attribute && state->stage != MESA_SHADER_VERTEX) {
var->type = glsl_type::error_type;
_mesa_glsl_error(loc, state,
"`attribute' variables may not be declared in the "
"%s shader",
_mesa_shader_stage_to_string(state->stage));
}
/* Disallow layout qualifiers which may only appear on layout declarations. */
if (qual->flags.q.prim_type) {
_mesa_glsl_error(loc, state,
"Primitive type may only be specified on GS input or output "
"layout declaration, not on variables.");
}
/* Section 6.1.1 (Function Calling Conventions) of the GLSL 1.10 spec says:
*
* "However, the const qualifier cannot be used with out or inout."
*
* The same section of the GLSL 4.40 spec further clarifies this saying:
*
* "The const qualifier cannot be used with out or inout, or a
* compile-time error results."
*/
if (is_parameter && qual->flags.q.constant && qual->flags.q.out) {
_mesa_glsl_error(loc, state,
"`const' may not be applied to `out' or `inout' "
"function parameters");
}
/* If there is no qualifier that changes the mode of the variable, leave
* the setting alone.
*/
assert(var->data.mode != ir_var_temporary);
if (qual->flags.q.in && qual->flags.q.out)
var->data.mode = is_parameter ? ir_var_function_inout : ir_var_shader_out;
else if (qual->flags.q.in)
var->data.mode = is_parameter ? ir_var_function_in : ir_var_shader_in;
else if (qual->flags.q.attribute
|| (qual->flags.q.varying && (state->stage == MESA_SHADER_FRAGMENT)))
var->data.mode = ir_var_shader_in;
else if (qual->flags.q.out)
var->data.mode = is_parameter ? ir_var_function_out : ir_var_shader_out;
else if (qual->flags.q.varying && (state->stage == MESA_SHADER_VERTEX))
var->data.mode = ir_var_shader_out;
else if (qual->flags.q.uniform)
var->data.mode = ir_var_uniform;
else if (qual->flags.q.buffer)
var->data.mode = ir_var_shader_storage;
else if (qual->flags.q.shared_storage)
var->data.mode = ir_var_shader_shared;
if (!is_parameter && state->has_framebuffer_fetch() &&
state->stage == MESA_SHADER_FRAGMENT) {
if (state->is_version(130, 300))
var->data.fb_fetch_output = qual->flags.q.in && qual->flags.q.out;
else
var->data.fb_fetch_output = (strcmp(var->name, "gl_LastFragData") == 0);
}
if (var->data.fb_fetch_output) {
var->data.assigned = true;
var->data.memory_coherent = !qual->flags.q.non_coherent;
/* From the EXT_shader_framebuffer_fetch spec:
*
* "It is an error to declare an inout fragment output not qualified
* with layout(noncoherent) if the GL_EXT_shader_framebuffer_fetch
* extension hasn't been enabled."
*/
if (var->data.memory_coherent &&
!state->EXT_shader_framebuffer_fetch_enable)
_mesa_glsl_error(loc, state,
"invalid declaration of framebuffer fetch output not "
"qualified with layout(noncoherent)");
} else {
/* From the EXT_shader_framebuffer_fetch spec:
*
* "Fragment outputs declared inout may specify the following layout
* qualifier: [...] noncoherent"
*/
if (qual->flags.q.non_coherent)
_mesa_glsl_error(loc, state,
"invalid layout(noncoherent) qualifier not part of "
"framebuffer fetch output declaration");
}
if (!is_parameter && is_varying_var(var, state->stage)) {
/* User-defined ins/outs are not permitted in compute shaders. */
if (state->stage == MESA_SHADER_COMPUTE) {
_mesa_glsl_error(loc, state,
"user-defined input and output variables are not "
"permitted in compute shaders");
}
/* This variable is being used to link data between shader stages (in
* pre-glsl-1.30 parlance, it's a "varying"). Check that it has a type
* that is allowed for such purposes.
*
* From page 25 (page 31 of the PDF) of the GLSL 1.10 spec:
*
* "The varying qualifier can be used only with the data types
* float, vec2, vec3, vec4, mat2, mat3, and mat4, or arrays of
* these."
*
* This was relaxed in GLSL version 1.30 and GLSL ES version 3.00. From
* page 31 (page 37 of the PDF) of the GLSL 1.30 spec:
*
* "Fragment inputs can only be signed and unsigned integers and
* integer vectors, float, floating-point vectors, matrices, or
* arrays of these. Structures cannot be input.
*
* Similar text exists in the section on vertex shader outputs.
*
* Similar text exists in the GLSL ES 3.00 spec, except that the GLSL ES
* 3.00 spec allows structs as well. Varying structs are also allowed
* in GLSL 1.50.
*
* From section 4.3.4 of the ARB_bindless_texture spec:
*
* "(modify third paragraph of the section to allow sampler and image
* types) ... Vertex shader inputs can only be float,
* single-precision floating-point scalars, single-precision
* floating-point vectors, matrices, signed and unsigned integers
* and integer vectors, sampler and image types."
*
* From section 4.3.6 of the ARB_bindless_texture spec:
*
* "Output variables can only be floating-point scalars,
* floating-point vectors, matrices, signed or unsigned integers or
* integer vectors, sampler or image types, or arrays or structures
* of any these."
*/
switch (var->type->without_array()->base_type) {
case GLSL_TYPE_FLOAT:
/* Ok in all GLSL versions */
break;
case GLSL_TYPE_UINT:
case GLSL_TYPE_INT:
if (state->is_version(130, 300) || state->EXT_gpu_shader4_enable)
break;
_mesa_glsl_error(loc, state,
"varying variables must be of base type float in %s",
state->get_version_string());
break;
case GLSL_TYPE_STRUCT:
if (state->is_version(150, 300))
break;
_mesa_glsl_error(loc, state,
"varying variables may not be of type struct");
break;
case GLSL_TYPE_DOUBLE:
case GLSL_TYPE_UINT64:
case GLSL_TYPE_INT64:
break;
case GLSL_TYPE_SAMPLER:
case GLSL_TYPE_IMAGE:
if (state->has_bindless())
break;
/* fallthrough */
default:
_mesa_glsl_error(loc, state, "illegal type for a varying variable");
break;
}
}
if (state->all_invariant && var->data.mode == ir_var_shader_out) {
var->data.explicit_invariant = true;
var->data.invariant = true;
}
var->data.interpolation =
interpret_interpolation_qualifier(qual, var->type,
(ir_variable_mode) var->data.mode,
state, loc);
/* Does the declaration use the deprecated 'attribute' or 'varying'
* keywords?
*/
const bool uses_deprecated_qualifier = qual->flags.q.attribute
|| qual->flags.q.varying;
/* Validate auxiliary storage qualifiers */
/* From section 4.3.4 of the GLSL 1.30 spec:
* "It is an error to use centroid in in a vertex shader."
*
* From section 4.3.4 of the GLSL ES 3.00 spec:
* "It is an error to use centroid in or interpolation qualifiers in
* a vertex shader input."
*/
/* Section 4.3.6 of the GLSL 1.30 specification states:
* "It is an error to use centroid out in a fragment shader."
*
* The GL_ARB_shading_language_420pack extension specification states:
* "It is an error to use auxiliary storage qualifiers or interpolation
* qualifiers on an output in a fragment shader."
*/
if (qual->flags.q.sample && (!is_varying_var(var, state->stage) || uses_deprecated_qualifier)) {
_mesa_glsl_error(loc, state,
"sample qualifier may only be used on `in` or `out` "
"variables between shader stages");
}
if (qual->flags.q.centroid && !is_varying_var(var, state->stage)) {
_mesa_glsl_error(loc, state,
"centroid qualifier may only be used with `in', "
"`out' or `varying' variables between shader stages");
}
if (qual->flags.q.shared_storage && state->stage != MESA_SHADER_COMPUTE) {
_mesa_glsl_error(loc, state,
"the shared storage qualifiers can only be used with "
"compute shaders");
}
apply_image_qualifier_to_variable(qual, var, state, loc);
}
/**
* Get the variable that is being redeclared by this declaration or if it
* does not exist, the current declared variable.
*
* Semantic checks to verify the validity of the redeclaration are also
* performed. If semantic checks fail, compilation error will be emitted via
* \c _mesa_glsl_error, but a non-\c NULL pointer will still be returned.
*
* \returns
* A pointer to an existing variable in the current scope if the declaration
* is a redeclaration, current variable otherwise. \c is_declared boolean
* will return \c true if the declaration is a redeclaration, \c false
* otherwise.
*/
static ir_variable *
get_variable_being_redeclared(ir_variable **var_ptr, YYLTYPE loc,
struct _mesa_glsl_parse_state *state,
bool allow_all_redeclarations,
bool *is_redeclaration)
{
ir_variable *var = *var_ptr;
/* Check if this declaration is actually a re-declaration, either to
* resize an array or add qualifiers to an existing variable.
*
* This is allowed for variables in the current scope, or when at
* global scope (for built-ins in the implicit outer scope).
*/
ir_variable *earlier = state->symbols->get_variable(var->name);
if (earlier == NULL ||
(state->current_function != NULL &&
!state->symbols->name_declared_this_scope(var->name))) {
*is_redeclaration = false;
return var;
}
*is_redeclaration = true;
if (earlier->data.how_declared == ir_var_declared_implicitly) {
/* Verify that the redeclaration of a built-in does not change the
* storage qualifier. There are a couple special cases.
*
* 1. Some built-in variables that are defined as 'in' in the
* specification are implemented as system values. Allow
* ir_var_system_value -> ir_var_shader_in.
*
* 2. gl_LastFragData is implemented as a ir_var_shader_out, but the
* specification requires that redeclarations omit any qualifier.
* Allow ir_var_shader_out -> ir_var_auto for this one variable.
*/
if (earlier->data.mode != var->data.mode &&
!(earlier->data.mode == ir_var_system_value &&
var->data.mode == ir_var_shader_in) &&
!(strcmp(var->name, "gl_LastFragData") == 0 &&
var->data.mode == ir_var_auto)) {
_mesa_glsl_error(&loc, state,
"redeclaration cannot change qualification of `%s'",
var->name);
}
}
/* From page 24 (page 30 of the PDF) of the GLSL 1.50 spec,
*
* "It is legal to declare an array without a size and then
* later re-declare the same name as an array of the same
* type and specify a size."
*/
if (earlier->type->is_unsized_array() && var->type->is_array()
&& (var->type->fields.array == earlier->type->fields.array)) {
const int size = var->type->array_size();
check_builtin_array_max_size(var->name, size, loc, state);
if ((size > 0) && (size <= earlier->data.max_array_access)) {
_mesa_glsl_error(& loc, state, "array size must be > %u due to "
"previous access",
earlier->data.max_array_access);
}
earlier->type = var->type;
delete var;
var = NULL;
*var_ptr = NULL;
} else if (earlier->type != var->type) {
_mesa_glsl_error(&loc, state,
"redeclaration of `%s' has incorrect type",
var->name);
} else if ((state->ARB_fragment_coord_conventions_enable ||
state->is_version(150, 0))
&& strcmp(var->name, "gl_FragCoord") == 0) {
/* Allow redeclaration of gl_FragCoord for ARB_fcc layout
* qualifiers.
*
* We don't really need to do anything here, just allow the
* redeclaration. Any error on the gl_FragCoord is handled on the ast
* level at apply_layout_qualifier_to_variable using the
* ast_type_qualifier and _mesa_glsl_parse_state, or later at
* linker.cpp.
*/
/* According to section 4.3.7 of the GLSL 1.30 spec,
* the following built-in varaibles can be redeclared with an
* interpolation qualifier:
* * gl_FrontColor
* * gl_BackColor
* * gl_FrontSecondaryColor
* * gl_BackSecondaryColor
* * gl_Color
* * gl_SecondaryColor
*/
} else if (state->is_version(130, 0)
&& (strcmp(var->name, "gl_FrontColor") == 0
|| strcmp(var->name, "gl_BackColor") == 0
|| strcmp(var->name, "gl_FrontSecondaryColor") == 0
|| strcmp(var->name, "gl_BackSecondaryColor") == 0
|| strcmp(var->name, "gl_Color") == 0
|| strcmp(var->name, "gl_SecondaryColor") == 0)) {
earlier->data.interpolation = var->data.interpolation;
/* Layout qualifiers for gl_FragDepth. */
} else if ((state->is_version(420, 0) ||
state->AMD_conservative_depth_enable ||
state->ARB_conservative_depth_enable)
&& strcmp(var->name, "gl_FragDepth") == 0) {
/** From the AMD_conservative_depth spec:
* Within any shader, the first redeclarations of gl_FragDepth
* must appear before any use of gl_FragDepth.
*/
if (earlier->data.used) {
_mesa_glsl_error(&loc, state,
"the first redeclaration of gl_FragDepth "
"must appear before any use of gl_FragDepth");
}
/* Prevent inconsistent redeclaration of depth layout qualifier. */
if (earlier->data.depth_layout != ir_depth_layout_none
&& earlier->data.depth_layout != var->data.depth_layout) {
_mesa_glsl_error(&loc, state,
"gl_FragDepth: depth layout is declared here "
"as '%s, but it was previously declared as "
"'%s'",
depth_layout_string(var->data.depth_layout),
depth_layout_string(earlier->data.depth_layout));
}
earlier->data.depth_layout = var->data.depth_layout;
} else if (state->has_framebuffer_fetch() &&
strcmp(var->name, "gl_LastFragData") == 0 &&
var->data.mode == ir_var_auto) {
/* According to the EXT_shader_framebuffer_fetch spec:
*
* "By default, gl_LastFragData is declared with the mediump precision
* qualifier. This can be changed by redeclaring the corresponding
* variables with the desired precision qualifier."
*
* "Fragment shaders may specify the following layout qualifier only for
* redeclaring the built-in gl_LastFragData array [...]: noncoherent"
*/
earlier->data.precision = var->data.precision;
earlier->data.memory_coherent = var->data.memory_coherent;
} else if (state->NV_viewport_array2_enable &&
strcmp(var->name, "gl_Layer") == 0 &&
earlier->data.how_declared == ir_var_declared_implicitly) {
/* No need to do anything, just allow it. Qualifier is stored in state */
} else if ((earlier->data.how_declared == ir_var_declared_implicitly &&
state->allow_builtin_variable_redeclaration) ||
allow_all_redeclarations) {
/* Allow verbatim redeclarations of built-in variables. Not explicitly
* valid, but some applications do it.
*/
} else {
_mesa_glsl_error(&loc, state, "`%s' redeclared", var->name);
}
return earlier;
}
/**
* Generate the IR for an initializer in a variable declaration
*/
static ir_rvalue *
process_initializer(ir_variable *var, ast_declaration *decl,
ast_fully_specified_type *type,
exec_list *initializer_instructions,
struct _mesa_glsl_parse_state *state)
{
void *mem_ctx = state;
ir_rvalue *result = NULL;
YYLTYPE initializer_loc = decl->initializer->get_location();
/* From page 24 (page 30 of the PDF) of the GLSL 1.10 spec:
*
* "All uniform variables are read-only and are initialized either
* directly by an application via API commands, or indirectly by
* OpenGL."
*/
if (var->data.mode == ir_var_uniform) {
state->check_version(120, 0, &initializer_loc,
"cannot initialize uniform %s",
var->name);
}
/* Section 4.3.7 "Buffer Variables" of the GLSL 4.30 spec:
*
* "Buffer variables cannot have initializers."
*/
if (var->data.mode == ir_var_shader_storage) {
_mesa_glsl_error(&initializer_loc, state,
"cannot initialize buffer variable %s",
var->name);
}
/* From section 4.1.7 of the GLSL 4.40 spec:
*
* "Opaque variables [...] are initialized only through the
* OpenGL API; they cannot be declared with an initializer in a
* shader."
*
* From section 4.1.7 of the ARB_bindless_texture spec:
*
* "Samplers may be declared as shader inputs and outputs, as uniform
* variables, as temporary variables, and as function parameters."
*
* From section 4.1.X of the ARB_bindless_texture spec:
*
* "Images may be declared as shader inputs and outputs, as uniform
* variables, as temporary variables, and as function parameters."
*/
if (var->type->contains_atomic() ||
(!state->has_bindless() && var->type->contains_opaque())) {
_mesa_glsl_error(&initializer_loc, state,
"cannot initialize %s variable %s",
var->name, state->has_bindless() ? "atomic" : "opaque");
}
if ((var->data.mode == ir_var_shader_in) && (state->current_function == NULL)) {
_mesa_glsl_error(&initializer_loc, state,
"cannot initialize %s shader input / %s %s",
_mesa_shader_stage_to_string(state->stage),
(state->stage == MESA_SHADER_VERTEX)
? "attribute" : "varying",
var->name);
}
if (var->data.mode == ir_var_shader_out && state->current_function == NULL) {
_mesa_glsl_error(&initializer_loc, state,
"cannot initialize %s shader output %s",
_mesa_shader_stage_to_string(state->stage),
var->name);
}
/* If the initializer is an ast_aggregate_initializer, recursively store
* type information from the LHS into it, so that its hir() function can do
* type checking.
*/
if (decl->initializer->oper == ast_aggregate)
_mesa_ast_set_aggregate_type(var->type, decl->initializer);
ir_dereference *const lhs = new(state) ir_dereference_variable(var);
ir_rvalue *rhs = decl->initializer->hir(initializer_instructions, state);
/* Calculate the constant value if this is a const or uniform
* declaration.
*
* Section 4.3 (Storage Qualifiers) of the GLSL ES 1.00.17 spec says:
*
* "Declarations of globals without a storage qualifier, or with
* just the const qualifier, may include initializers, in which case
* they will be initialized before the first line of main() is
* executed. Such initializers must be a constant expression."
*
* The same section of the GLSL ES 3.00.4 spec has similar language.
*/
if (type->qualifier.flags.q.constant
|| type->qualifier.flags.q.uniform
|| (state->es_shader && state->current_function == NULL)) {
ir_rvalue *new_rhs = validate_assignment(state, initializer_loc,
lhs, rhs, true);
if (new_rhs != NULL) {
rhs = new_rhs;
/* Section 4.3.3 (Constant Expressions) of the GLSL ES 3.00.4 spec
* says:
*
* "A constant expression is one of
*
* ...
*
* - an expression formed by an operator on operands that are
* all constant expressions, including getting an element of
* a constant array, or a field of a constant structure, or
* components of a constant vector. However, the sequence
* operator ( , ) and the assignment operators ( =, +=, ...)
* are not included in the operators that can create a
* constant expression."
*
* Section 12.43 (Sequence operator and constant expressions) says:
*
* "Should the following construct be allowed?
*
* float a[2,3];
*
* The expression within the brackets uses the sequence operator
* (',') and returns the integer 3 so the construct is declaring
* a single-dimensional array of size 3. In some languages, the
* construct declares a two-dimensional array. It would be
* preferable to make this construct illegal to avoid confusion.
*
* One possibility is to change the definition of the sequence
* operator so that it does not return a constant-expression and
* hence cannot be used to declare an array size.
*
* RESOLUTION: The result of a sequence operator is not a
* constant-expression."
*
* Section 4.3.3 (Constant Expressions) of the GLSL 4.30.9 spec
* contains language almost identical to the section 4.3.3 in the
* GLSL ES 3.00.4 spec. This is a new limitation for these GLSL
* versions.
*/
ir_constant *constant_value =
rhs->constant_expression_value(mem_ctx);
if (!constant_value ||
(state->is_version(430, 300) &&
decl->initializer->has_sequence_subexpression())) {
const char *const variable_mode =
(type->qualifier.flags.q.constant)
? "const"
: ((type->qualifier.flags.q.uniform) ? "uniform" : "global");
/* If ARB_shading_language_420pack is enabled, initializers of
* const-qualified local variables do not have to be constant
* expressions. Const-qualified global variables must still be
* initialized with constant expressions.
*/
if (!state->has_420pack()
|| state->current_function == NULL) {
_mesa_glsl_error(& initializer_loc, state,
"initializer of %s variable `%s' must be a "
"constant expression",
variable_mode,
decl->identifier);
if (var->type->is_numeric()) {
/* Reduce cascading errors. */
var->constant_value = type->qualifier.flags.q.constant
? ir_constant::zero(state, var->type) : NULL;
}
}
} else {
rhs = constant_value;
var->constant_value = type->qualifier.flags.q.constant
? constant_value : NULL;
}
} else {
if (var->type->is_numeric()) {
/* Reduce cascading errors. */
rhs = var->constant_value = type->qualifier.flags.q.constant
? ir_constant::zero(state, var->type) : NULL;
}
}
}
if (rhs && !rhs->type->is_error()) {
bool temp = var->data.read_only;
if (type->qualifier.flags.q.constant)
var->data.read_only = false;
/* Never emit code to initialize a uniform.
*/
const glsl_type *initializer_type;
bool error_emitted = false;
if (!type->qualifier.flags.q.uniform) {
error_emitted =
do_assignment(initializer_instructions, state,
NULL, lhs, rhs,
&result, true, true,
type->get_location());
initializer_type = result->type;
} else
initializer_type = rhs->type;
if (!error_emitted) {
var->constant_initializer = rhs->constant_expression_value(mem_ctx);
var->data.has_initializer = true;
/* If the declared variable is an unsized array, it must inherrit
* its full type from the initializer. A declaration such as
*
* uniform float a[] = float[](1.0, 2.0, 3.0, 3.0);
*
* becomes
*
* uniform float a[4] = float[](1.0, 2.0, 3.0, 3.0);
*
* The assignment generated in the if-statement (below) will also
* automatically handle this case for non-uniforms.
*
* If the declared variable is not an array, the types must
* already match exactly. As a result, the type assignment
* here can be done unconditionally. For non-uniforms the call
* to do_assignment can change the type of the initializer (via
* the implicit conversion rules). For uniforms the initializer
* must be a constant expression, and the type of that expression
* was validated above.
*/
var->type = initializer_type;
}
var->data.read_only = temp;
}
return result;
}
static void
validate_layout_qualifier_vertex_count(struct _mesa_glsl_parse_state *state,
YYLTYPE loc, ir_variable *var,
unsigned num_vertices,
unsigned *size,
const char *var_category)
{
if (var->type->is_unsized_array()) {
/* Section 4.3.8.1 (Input Layout Qualifiers) of the GLSL 1.50 spec says:
*
* All geometry shader input unsized array declarations will be
* sized by an earlier input layout qualifier, when present, as per
* the following table.
*
* Followed by a table mapping each allowed input layout qualifier to
* the corresponding input length.
*
* Similarly for tessellation control shader outputs.
*/
if (num_vertices != 0)
var->type = glsl_type::get_array_instance(var->type->fields.array,
num_vertices);
} else {
/* Section 4.3.8.1 (Input Layout Qualifiers) of the GLSL 1.50 spec
* includes the following examples of compile-time errors:
*
* // code sequence within one shader...
* in vec4 Color1[]; // size unknown
* ...Color1.length()...// illegal, length() unknown
* in vec4 Color2[2]; // size is 2
* ...Color1.length()...// illegal, Color1 still has no size
* in vec4 Color3[3]; // illegal, input sizes are inconsistent
* layout(lines) in; // legal, input size is 2, matching
* in vec4 Color4[3]; // illegal, contradicts layout
* ...
*
* To detect the case illustrated by Color3, we verify that the size of
* an explicitly-sized array matches the size of any previously declared
* explicitly-sized array. To detect the case illustrated by Color4, we
* verify that the size of an explicitly-sized array is consistent with
* any previously declared input layout.
*/
if (num_vertices != 0 && var->type->length != num_vertices) {
_mesa_glsl_error(&loc, state,
"%s size contradicts previously declared layout "
"(size is %u, but layout requires a size of %u)",
var_category, var->type->length, num_vertices);
} else if (*size != 0 && var->type->length != *size) {
_mesa_glsl_error(&loc, state,
"%s sizes are inconsistent (size is %u, but a "
"previous declaration has size %u)",
var_category, var->type->length, *size);
} else {
*size = var->type->length;
}
}
}
static void
handle_tess_ctrl_shader_output_decl(struct _mesa_glsl_parse_state *state,
YYLTYPE loc, ir_variable *var)
{
unsigned num_vertices = 0;
if (state->tcs_output_vertices_specified) {
if (!state->out_qualifier->vertices->
process_qualifier_constant(state, "vertices",
&num_vertices, false)) {
return;
}
if (num_vertices > state->Const.MaxPatchVertices) {
_mesa_glsl_error(&loc, state, "vertices (%d) exceeds "
"GL_MAX_PATCH_VERTICES", num_vertices);
return;
}
}
if (!var->type->is_array() && !var->data.patch) {
_mesa_glsl_error(&loc, state,
"tessellation control shader outputs must be arrays");
/* To avoid cascading failures, short circuit the checks below. */
return;
}
if (var->data.patch)
return;
validate_layout_qualifier_vertex_count(state, loc, var, num_vertices,
&state->tcs_output_size,
"tessellation control shader output");
}
/**
* Do additional processing necessary for tessellation control/evaluation shader
* input declarations. This covers both interface block arrays and bare input
* variables.
*/
static void
handle_tess_shader_input_decl(struct _mesa_glsl_parse_state *state,
YYLTYPE loc, ir_variable *var)
{
if (!var->type->is_array() && !var->data.patch) {
_mesa_glsl_error(&loc, state,
"per-vertex tessellation shader inputs must be arrays");
/* Avoid cascading failures. */
return;
}
if (var->data.patch)
return;
/* The ARB_tessellation_shader spec says:
*
* "Declaring an array size is optional. If no size is specified, it
* will be taken from the implementation-dependent maximum patch size
* (gl_MaxPatchVertices). If a size is specified, it must match the
* maximum patch size; otherwise, a compile or link error will occur."
*
* This text appears twice, once for TCS inputs, and again for TES inputs.
*/
if (var->type->is_unsized_array()) {
var->type = glsl_type::get_array_instance(var->type->fields.array,
state->Const.MaxPatchVertices);
} else if (var->type->length != state->Const.MaxPatchVertices) {
_mesa_glsl_error(&loc, state,
"per-vertex tessellation shader input arrays must be "
"sized to gl_MaxPatchVertices (%d).",
state->Const.MaxPatchVertices);
}
}
/**
* Do additional processing necessary for geometry shader input declarations
* (this covers both interface blocks arrays and bare input variables).
*/
static void
handle_geometry_shader_input_decl(struct _mesa_glsl_parse_state *state,
YYLTYPE loc, ir_variable *var)
{
unsigned num_vertices = 0;
if (state->gs_input_prim_type_specified) {
num_vertices = vertices_per_prim(state->in_qualifier->prim_type);
}
/* Geometry shader input variables must be arrays. Caller should have
* reported an error for this.
*/
if (!var->type->is_array()) {
assert(state->error);
/* To avoid cascading failures, short circuit the checks below. */
return;
}
validate_layout_qualifier_vertex_count(state, loc, var, num_vertices,
&state->gs_input_size,
"geometry shader input");
}
static void
validate_identifier(const char *identifier, YYLTYPE loc,
struct _mesa_glsl_parse_state *state)
{
/* From page 15 (page 21 of the PDF) of the GLSL 1.10 spec,
*
* "Identifiers starting with "gl_" are reserved for use by
* OpenGL, and may not be declared in a shader as either a
* variable or a function."
*/
if (is_gl_identifier(identifier)) {
_mesa_glsl_error(&loc, state,
"identifier `%s' uses reserved `gl_' prefix",
identifier);
} else if (strstr(identifier, "__")) {
/* From page 14 (page 20 of the PDF) of the GLSL 1.10
* spec:
*
* "In addition, all identifiers containing two
* consecutive underscores (__) are reserved as
* possible future keywords."
*
* The intention is that names containing __ are reserved for internal
* use by the implementation, and names prefixed with GL_ are reserved
* for use by Khronos. Names simply containing __ are dangerous to use,
* but should be allowed.
*
* A future version of the GLSL specification will clarify this.
*/
_mesa_glsl_warning(&loc, state,
"identifier `%s' uses reserved `__' string",
identifier);
}
}
ir_rvalue *
ast_declarator_list::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
const struct glsl_type *decl_type;
const char *type_name = NULL;
ir_rvalue *result = NULL;
YYLTYPE loc = this->get_location();
/* From page 46 (page 52 of the PDF) of the GLSL 1.50 spec:
*
* "To ensure that a particular output variable is invariant, it is
* necessary to use the invariant qualifier. It can either be used to
* qualify a previously declared variable as being invariant
*
* invariant gl_Position; // make existing gl_Position be invariant"
*
* In these cases the parser will set the 'invariant' flag in the declarator
* list, and the type will be NULL.
*/
if (this->invariant) {
assert(this->type == NULL);
if (state->current_function != NULL) {
_mesa_glsl_error(& loc, state,
"all uses of `invariant' keyword must be at global "
"scope");
}
foreach_list_typed (ast_declaration, decl, link, &this->declarations) {
assert(decl->array_specifier == NULL);
assert(decl->initializer == NULL);
ir_variable *const earlier =
state->symbols->get_variable(decl->identifier);
if (earlier == NULL) {
_mesa_glsl_error(& loc, state,
"undeclared variable `%s' cannot be marked "
"invariant", decl->identifier);
} else if (!is_allowed_invariant(earlier, state)) {
_mesa_glsl_error(&loc, state,
"`%s' cannot be marked invariant; interfaces between "
"shader stages only.", decl->identifier);
} else if (earlier->data.used) {
_mesa_glsl_error(& loc, state,
"variable `%s' may not be redeclared "
"`invariant' after being used",
earlier->name);
} else {
earlier->data.explicit_invariant = true;
earlier->data.invariant = true;
}
}
/* Invariant redeclarations do not have r-values.
*/
return NULL;
}
if (this->precise) {
assert(this->type == NULL);
foreach_list_typed (ast_declaration, decl, link, &this->declarations) {
assert(decl->array_specifier == NULL);
assert(decl->initializer == NULL);
ir_variable *const earlier =
state->symbols->get_variable(decl->identifier);
if (earlier == NULL) {
_mesa_glsl_error(& loc, state,
"undeclared variable `%s' cannot be marked "
"precise", decl->identifier);
} else if (state->current_function != NULL &&
!state->symbols->name_declared_this_scope(decl->identifier)) {
/* Note: we have to check if we're in a function, since
* builtins are treated as having come from another scope.
*/
_mesa_glsl_error(& loc, state,
"variable `%s' from an outer scope may not be "
"redeclared `precise' in this scope",
earlier->name);
} else if (earlier->data.used) {
_mesa_glsl_error(& loc, state,
"variable `%s' may not be redeclared "
"`precise' after being used",
earlier->name);
} else {
earlier->data.precise = true;
}
}
/* Precise redeclarations do not have r-values either. */
return NULL;
}
assert(this->type != NULL);
assert(!this->invariant);
assert(!this->precise);
/* GL_EXT_shader_image_load_store base type uses GLSL_TYPE_VOID as a special value to
* indicate that it needs to be updated later (see glsl_parser.yy).
* This is done here, based on the layout qualifier and the type of the image var
*/
if (this->type->qualifier.flags.q.explicit_image_format &&
this->type->specifier->type->is_image() &&
this->type->qualifier.image_base_type == GLSL_TYPE_VOID) {
/* "The ARB_shader_image_load_store says:
* If both extensions are enabled in the shading language, the "size*" layout
* qualifiers are treated as format qualifiers, and are mapped to equivalent
* format qualifiers in the table below, according to the type of image
* variable.
* image* iimage* uimage*
* -------- -------- --------
* size1x8 n/a r8i r8ui
* size1x16 r16f r16i r16ui
* size1x32 r32f r32i r32ui
* size2x32 rg32f rg32i rg32ui
* size4x32 rgba32f rgba32i rgba32ui"
*/
if (strncmp(this->type->specifier->type_name, "image", strlen("image")) == 0) {
switch (this->type->qualifier.image_format) {
case PIPE_FORMAT_R8_SINT:
/* No valid qualifier in this case, driver will need to look at
* the underlying image's format (just like no qualifier being
* present).
*/
this->type->qualifier.image_format = PIPE_FORMAT_NONE;
break;
case PIPE_FORMAT_R16_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R16_FLOAT;
break;
case PIPE_FORMAT_R32_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R32_FLOAT;
break;
case PIPE_FORMAT_R32G32_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R32G32_FLOAT;
break;
case PIPE_FORMAT_R32G32B32A32_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R32G32B32A32_FLOAT;
break;
default:
unreachable("Unknown image format");
}
this->type->qualifier.image_base_type = GLSL_TYPE_FLOAT;
} else if (strncmp(this->type->specifier->type_name, "uimage", strlen("uimage")) == 0) {
switch (this->type->qualifier.image_format) {
case PIPE_FORMAT_R8_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R8_UINT;
break;
case PIPE_FORMAT_R16_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R16_UINT;
break;
case PIPE_FORMAT_R32_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R32_UINT;
break;
case PIPE_FORMAT_R32G32_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R32G32_UINT;
break;
case PIPE_FORMAT_R32G32B32A32_SINT:
this->type->qualifier.image_format = PIPE_FORMAT_R32G32B32A32_UINT;
break;
default:
unreachable("Unknown image format");
}
this->type->qualifier.image_base_type = GLSL_TYPE_UINT;
} else if (strncmp(this->type->specifier->type_name, "iimage", strlen("iimage")) == 0) {
this->type->qualifier.image_base_type = GLSL_TYPE_INT;
} else {
assert(false);
}
}
/* The type specifier may contain a structure definition. Process that
* before any of the variable declarations.
*/
(void) this->type->specifier->hir(instructions, state);
decl_type = this->type->glsl_type(& type_name, state);
/* Section 4.3.7 "Buffer Variables" of the GLSL 4.30 spec:
* "Buffer variables may only be declared inside interface blocks
* (section 4.3.9 “Interface Blocks”), which are then referred to as
* shader storage blocks. It is a compile-time error to declare buffer
* variables at global scope (outside a block)."
*/
if (type->qualifier.flags.q.buffer && !decl_type->is_interface()) {
_mesa_glsl_error(&loc, state,
"buffer variables cannot be declared outside "
"interface blocks");
}
/* An offset-qualified atomic counter declaration sets the default
* offset for the next declaration within the same atomic counter
* buffer.
*/
if (decl_type && decl_type->contains_atomic()) {
if (type->qualifier.flags.q.explicit_binding &&
type->qualifier.flags.q.explicit_offset) {
unsigned qual_binding;
unsigned qual_offset;
if (process_qualifier_constant(state, &loc, "binding",
type->qualifier.binding,
&qual_binding)
&& process_qualifier_constant(state, &loc, "offset",
type->qualifier.offset,
&qual_offset)) {
if (qual_binding < ARRAY_SIZE(state->atomic_counter_offsets))
state->atomic_counter_offsets[qual_binding] = qual_offset;
}
}
ast_type_qualifier allowed_atomic_qual_mask;
allowed_atomic_qual_mask.flags.i = 0;
allowed_atomic_qual_mask.flags.q.explicit_binding = 1;
allowed_atomic_qual_mask.flags.q.explicit_offset = 1;
allowed_atomic_qual_mask.flags.q.uniform = 1;
type->qualifier.validate_flags(&loc, state, allowed_atomic_qual_mask,
"invalid layout qualifier for",
"atomic_uint");
}
if (this->declarations.is_empty()) {
/* If there is no structure involved in the program text, there are two
* possible scenarios:
*
* - The program text contained something like 'vec4;'. This is an
* empty declaration. It is valid but weird. Emit a warning.
*
* - The program text contained something like 'S;' and 'S' is not the
* name of a known structure type. This is both invalid and weird.
* Emit an error.
*
* - The program text contained something like 'mediump float;'
* when the programmer probably meant 'precision mediump
* float;' Emit a warning with a description of what they
* probably meant to do.
*
* Note that if decl_type is NULL and there is a structure involved,
* there must have been some sort of error with the structure. In this
* case we assume that an error was already generated on this line of
* code for the structure. There is no need to generate an additional,
* confusing error.
*/
assert(this->type->specifier->structure == NULL || decl_type != NULL
|| state->error);
if (decl_type == NULL) {
_mesa_glsl_error(&loc, state,
"invalid type `%s' in empty declaration",
type_name);
} else {
if (decl_type->is_array()) {
/* From Section 13.22 (Array Declarations) of the GLSL ES 3.2
* spec:
*
* "... any declaration that leaves the size undefined is
* disallowed as this would add complexity and there are no
* use-cases."
*/
if (state->es_shader && decl_type->is_unsized_array()) {
_mesa_glsl_error(&loc, state, "array size must be explicitly "
"or implicitly defined");
}
/* From Section 4.12 (Empty Declarations) of the GLSL 4.5 spec:
*
* "The combinations of types and qualifiers that cause
* compile-time or link-time errors are the same whether or not
* the declaration is empty."
*/
validate_array_dimensions(decl_type, state, &loc);
}
if (decl_type->is_atomic_uint()) {
/* Empty atomic counter declarations are allowed and useful
* to set the default offset qualifier.
*/
return NULL;
} else if (this->type->qualifier.precision != ast_precision_none) {
if (this->type->specifier->structure != NULL) {
_mesa_glsl_error(&loc, state,
"precision qualifiers can't be applied "
"to structures");
} else {
static const char *const precision_names[] = {
"highp",
"highp",
"mediump",
"lowp"
};
_mesa_glsl_warning(&loc, state,
"empty declaration with precision "
"qualifier, to set the default precision, "
"use `precision %s %s;'",
precision_names[this->type->
qualifier.precision],
type_name);
}
} else if (this->type->specifier->structure == NULL) {
_mesa_glsl_warning(&loc, state, "empty declaration");
}
}
}
foreach_list_typed (ast_declaration, decl, link, &this->declarations) {
const struct glsl_type *var_type;
ir_variable *var;
const char *identifier = decl->identifier;
/* FINISHME: Emit a warning if a variable declaration shadows a
* FINISHME: declaration at a higher scope.
*/
if ((decl_type == NULL) || decl_type->is_void()) {
if (type_name != NULL) {
_mesa_glsl_error(& loc, state,
"invalid type `%s' in declaration of `%s'",
type_name, decl->identifier);
} else {
_mesa_glsl_error(& loc, state,
"invalid type in declaration of `%s'",
decl->identifier);
}
continue;
}
if (this->type->qualifier.is_subroutine_decl()) {
const glsl_type *t;
const char *name;
t = state->symbols->get_type(this->type->specifier->type_name);
if (!t)
_mesa_glsl_error(& loc, state,
"invalid type in declaration of `%s'",
decl->identifier);
name = ralloc_asprintf(ctx, "%s_%s", _mesa_shader_stage_to_subroutine_prefix(state->stage), decl->identifier);
identifier = name;
}
var_type = process_array_type(&loc, decl_type, decl->array_specifier,
state);
var = new(ctx) ir_variable(var_type, identifier, ir_var_auto);
/* The 'varying in' and 'varying out' qualifiers can only be used with
* ARB_geometry_shader4 and EXT_geometry_shader4, which we don't support
* yet.
*/
if (this->type->qualifier.flags.q.varying) {
if (this->type->qualifier.flags.q.in) {
_mesa_glsl_error(& loc, state,
"`varying in' qualifier in declaration of "
"`%s' only valid for geometry shaders using "
"ARB_geometry_shader4 or EXT_geometry_shader4",
decl->identifier);
} else if (this->type->qualifier.flags.q.out) {
_mesa_glsl_error(& loc, state,
"`varying out' qualifier in declaration of "
"`%s' only valid for geometry shaders using "
"ARB_geometry_shader4 or EXT_geometry_shader4",
decl->identifier);
}
}
/* From page 22 (page 28 of the PDF) of the GLSL 1.10 specification;
*
* "Global variables can only use the qualifiers const,
* attribute, uniform, or varying. Only one may be
* specified.
*
* Local variables can only use the qualifier const."
*
* This is relaxed in GLSL 1.30 and GLSL ES 3.00. It is also relaxed by
* any extension that adds the 'layout' keyword.
*/
if (!state->is_version(130, 300)
&& !state->has_explicit_attrib_location()
&& !state->has_separate_shader_objects()
&& !state->ARB_fragment_coord_conventions_enable) {
/* GL_EXT_gpu_shader4 only allows "varying out" on fragment shader
* outputs. (the varying flag is not set by the parser)
*/
if (this->type->qualifier.flags.q.out &&
(!state->EXT_gpu_shader4_enable ||
state->stage != MESA_SHADER_FRAGMENT)) {
_mesa_glsl_error(& loc, state,
"`out' qualifier in declaration of `%s' "
"only valid for function parameters in %s",
decl->identifier, state->get_version_string());
}
if (this->type->qualifier.flags.q.in) {
_mesa_glsl_error(& loc, state,
"`in' qualifier in declaration of `%s' "
"only valid for function parameters in %s",
decl->identifier, state->get_version_string());
}
/* FINISHME: Test for other invalid qualifiers. */
}
apply_type_qualifier_to_variable(& this->type->qualifier, var, state,
& loc, false);
apply_layout_qualifier_to_variable(&this->type->qualifier, var, state,
&loc);
if ((var->data.mode == ir_var_auto || var->data.mode == ir_var_temporary
|| var->data.mode == ir_var_shader_out)
&& (var->type->is_numeric() || var->type->is_boolean())
&& state->zero_init) {
const ir_constant_data data = { { 0 } };
var->data.has_initializer = true;
var->constant_initializer = new(var) ir_constant(var->type, &data);
}
if (this->type->qualifier.flags.q.invariant) {
if (!is_allowed_invariant(var, state)) {
_mesa_glsl_error(&loc, state,
"`%s' cannot be marked invariant; interfaces between "
"shader stages only", var->name);
}
}
if (state->current_function != NULL) {
const char *mode = NULL;
const char *extra = "";
/* There is no need to check for 'inout' here because the parser will
* only allow that in function parameter lists.
*/
if (this->type->qualifier.flags.q.attribute) {
mode = "attribute";
} else if (this->type->qualifier.is_subroutine_decl()) {
mode = "subroutine uniform";
} else if (this->type->qualifier.flags.q.uniform) {
mode = "uniform";
} else if (this->type->qualifier.flags.q.varying) {
mode = "varying";
} else if (this->type->qualifier.flags.q.in) {
mode = "in";
extra = " or in function parameter list";
} else if (this->type->qualifier.flags.q.out) {
mode = "out";
extra = " or in function parameter list";
}
if (mode) {
_mesa_glsl_error(& loc, state,
"%s variable `%s' must be declared at "
"global scope%s",
mode, var->name, extra);
}
} else if (var->data.mode == ir_var_shader_in) {
var->data.read_only = true;
if (state->stage == MESA_SHADER_VERTEX) {
bool error_emitted = false;
/* From page 31 (page 37 of the PDF) of the GLSL 1.50 spec:
*
* "Vertex shader inputs can only be float, floating-point
* vectors, matrices, signed and unsigned integers and integer
* vectors. Vertex shader inputs can also form arrays of these
* types, but not structures."
*
* From page 31 (page 27 of the PDF) of the GLSL 1.30 spec:
*
* "Vertex shader inputs can only be float, floating-point
* vectors, matrices, signed and unsigned integers and integer
* vectors. They cannot be arrays or structures."
*
* From page 23 (page 29 of the PDF) of the GLSL 1.20 spec:
*
* "The attribute qualifier can be used only with float,
* floating-point vectors, and matrices. Attribute variables
* cannot be declared as arrays or structures."
*
* From page 33 (page 39 of the PDF) of the GLSL ES 3.00 spec:
*
* "Vertex shader inputs can only be float, floating-point
* vectors, matrices, signed and unsigned integers and integer
* vectors. Vertex shader inputs cannot be arrays or
* structures."
*
* From section 4.3.4 of the ARB_bindless_texture spec:
*
* "(modify third paragraph of the section to allow sampler and
* image types) ... Vertex shader inputs can only be float,
* single-precision floating-point scalars, single-precision
* floating-point vectors, matrices, signed and unsigned
* integers and integer vectors, sampler and image types."
*/
const glsl_type *check_type = var->type->without_array();
switch (check_type->base_type) {
case GLSL_TYPE_FLOAT:
break;
case GLSL_TYPE_UINT64:
case GLSL_TYPE_INT64:
break;
case GLSL_TYPE_UINT:
case GLSL_TYPE_INT:
if (state->is_version(120, 300) || state->EXT_gpu_shader4_enable)
break;
case GLSL_TYPE_DOUBLE:
if (check_type->is_double() && (state->is_version(410, 0) || state->ARB_vertex_attrib_64bit_enable))
break;
case GLSL_TYPE_SAMPLER:
if (check_type->is_sampler() && state->has_bindless())
break;
case GLSL_TYPE_IMAGE:
if (check_type->is_image() && state->has_bindless())
break;
/* FALLTHROUGH */
default:
_mesa_glsl_error(& loc, state,
"vertex shader input / attribute cannot have "
"type %s`%s'",
var->type->is_array() ? "array of " : "",
check_type->name);
error_emitted = true;
}
if (!error_emitted && var->type->is_array() &&
!state->check_version(150, 0, &loc,
"vertex shader input / attribute "
"cannot have array type")) {
error_emitted = true;
}
} else if (state->stage == MESA_SHADER_GEOMETRY) {
/* From section 4.3.4 (Inputs) of the GLSL 1.50 spec:
*
* Geometry shader input variables get the per-vertex values
* written out by vertex shader output variables of the same
* names. Since a geometry shader operates on a set of
* vertices, each input varying variable (or input block, see
* interface blocks below) needs to be declared as an array.
*/
if (!var->type->is_array()) {
_mesa_glsl_error(&loc, state,
"geometry shader inputs must be arrays");
}
handle_geometry_shader_input_decl(state, loc, var);
} else if (state->stage == MESA_SHADER_FRAGMENT) {
/* From section 4.3.4 (Input Variables) of the GLSL ES 3.10 spec:
*
* It is a compile-time error to declare a fragment shader
* input with, or that contains, any of the following types:
*
* * A boolean type
* * An opaque type
* * An array of arrays
* * An array of structures
* * A structure containing an array
* * A structure containing a structure
*/
if (state->es_shader) {
const glsl_type *check_type = var->type->without_array();
if (check_type->is_boolean() ||
check_type->contains_opaque()) {
_mesa_glsl_error(&loc, state,
"fragment shader input cannot have type %s",
check_type->name);
}
if (var->type->is_array() &&
var->type->fields.array->is_array()) {
_mesa_glsl_error(&loc, state,
"%s shader output "
"cannot have an array of arrays",
_mesa_shader_stage_to_string(state->stage));
}
if (var->type->is_array() &&
var->type->fields.array->is_struct()) {
_mesa_glsl_error(&loc, state,
"fragment shader input "
"cannot have an array of structs");
}
if (var->type->is_struct()) {
for (unsigned i = 0; i < var->type->length; i++) {
if (var->type->fields.structure[i].type->is_array() ||
var->type->fields.structure[i].type->is_struct())
_mesa_glsl_error(&loc, state,
"fragment shader input cannot have "
"a struct that contains an "
"array or struct");
}
}
}
} else if (state->stage == MESA_SHADER_TESS_CTRL ||
state->stage == MESA_SHADER_TESS_EVAL) {
handle_tess_shader_input_decl(state, loc, var);
}
} else if (var->data.mode == ir_var_shader_out) {
const glsl_type *check_type = var->type->without_array();
/* From section 4.3.6 (Output variables) of the GLSL 4.40 spec:
*
* It is a compile-time error to declare a fragment shader output
* that contains any of the following:
*
* * A Boolean type (bool, bvec2 ...)
* * A double-precision scalar or vector (double, dvec2 ...)
* * An opaque type
* * Any matrix type
* * A structure
*/
if (state->stage == MESA_SHADER_FRAGMENT) {
if (check_type->is_struct() || check_type->is_matrix())
_mesa_glsl_error(&loc, state,
"fragment shader output "
"cannot have struct or matrix type");
switch (check_type->base_type) {
case GLSL_TYPE_UINT:
case GLSL_TYPE_INT:
case GLSL_TYPE_FLOAT:
break;
default:
_mesa_glsl_error(&loc, state,
"fragment shader output cannot have "
"type %s", check_type->name);
}
}
/* From section 4.3.6 (Output Variables) of the GLSL ES 3.10 spec:
*
* It is a compile-time error to declare a vertex shader output
* with, or that contains, any of the following types:
*
* * A boolean type
* * An opaque type
* * An array of arrays
* * An array of structures
* * A structure containing an array
* * A structure containing a structure
*
* It is a compile-time error to declare a fragment shader output
* with, or that contains, any of the following types:
*
* * A boolean type
* * An opaque type
* * A matrix
* * A structure
* * An array of array
*
* ES 3.20 updates this to apply to tessellation and geometry shaders
* as well. Because there are per-vertex arrays in the new stages,
* it strikes the "array of..." rules and replaces them with these:
*
* * For per-vertex-arrayed variables (applies to tessellation
* control, tessellation evaluation and geometry shaders):
*
* * Per-vertex-arrayed arrays of arrays
* * Per-vertex-arrayed arrays of structures
*
* * For non-per-vertex-arrayed variables:
*
* * An array of arrays
* * An array of structures
*
* which basically says to unwrap the per-vertex aspect and apply
* the old rules.
*/
if (state->es_shader) {
if (var->type->is_array() &&
var->type->fields.array->is_array()) {
_mesa_glsl_error(&loc, state,
"%s shader output "
"cannot have an array of arrays",
_mesa_shader_stage_to_string(state->stage));
}
if (state->stage <= MESA_SHADER_GEOMETRY) {
const glsl_type *type = var->type;
if (state->stage == MESA_SHADER_TESS_CTRL &&
!var->data.patch && var->type->is_array()) {
type = var->type->fields.array;
}
if (type->is_array() && type->fields.array->is_struct()) {
_mesa_glsl_error(&loc, state,
"%s shader output cannot have "
"an array of structs",
_mesa_shader_stage_to_string(state->stage));
}
if (type->is_struct()) {
for (unsigned i = 0; i < type->length; i++) {
if (type->fields.structure[i].type->is_array() ||
type->fields.structure[i].type->is_struct())
_mesa_glsl_error(&loc, state,
"%s shader output cannot have a "
"struct that contains an "
"array or struct",
_mesa_shader_stage_to_string(state->stage));
}
}
}
}
if (state->stage == MESA_SHADER_TESS_CTRL) {
handle_tess_ctrl_shader_output_decl(state, loc, var);
}
} else if (var->type->contains_subroutine()) {
/* declare subroutine uniforms as hidden */
var->data.how_declared = ir_var_hidden;
}
/* From section 4.3.4 of the GLSL 4.00 spec:
* "Input variables may not be declared using the patch in qualifier
* in tessellation control or geometry shaders."
*
* From section 4.3.6 of the GLSL 4.00 spec:
* "It is an error to use patch out in a vertex, tessellation
* evaluation, or geometry shader."
*
* This doesn't explicitly forbid using them in a fragment shader, but
* that's probably just an oversight.
*/
if (state->stage != MESA_SHADER_TESS_EVAL
&& this->type->qualifier.flags.q.patch
&& this->type->qualifier.flags.q.in) {
_mesa_glsl_error(&loc, state, "'patch in' can only be used in a "
"tessellation evaluation shader");
}
if (state->stage != MESA_SHADER_TESS_CTRL
&& this->type->qualifier.flags.q.patch
&& this->type->qualifier.flags.q.out) {
_mesa_glsl_error(&loc, state, "'patch out' can only be used in a "
"tessellation control shader");
}
/* Precision qualifiers exists only in GLSL versions 1.00 and >= 1.30.
*/
if (this->type->qualifier.precision != ast_precision_none) {
state->check_precision_qualifiers_allowed(&loc);
}
if (this->type->qualifier.precision != ast_precision_none &&
!precision_qualifier_allowed(var->type)) {
_mesa_glsl_error(&loc, state,
"precision qualifiers apply only to floating point"
", integer and opaque types");
}
/* From section 4.1.7 of the GLSL 4.40 spec:
*
* "[Opaque types] can only be declared as function
* parameters or uniform-qualified variables."
*
* From section 4.1.7 of the ARB_bindless_texture spec:
*
* "Samplers may be declared as shader inputs and outputs, as uniform
* variables, as temporary variables, and as function parameters."
*
* From section 4.1.X of the ARB_bindless_texture spec:
*
* "Images may be declared as shader inputs and outputs, as uniform
* variables, as temporary variables, and as function parameters."
*/
if (!this->type->qualifier.flags.q.uniform &&
(var_type->contains_atomic() ||
(!state->has_bindless() && var_type->contains_opaque()))) {
_mesa_glsl_error(&loc, state,
"%s variables must be declared uniform",
state->has_bindless() ? "atomic" : "opaque");
}
/* Process the initializer and add its instructions to a temporary
* list. This list will be added to the instruction stream (below) after
* the declaration is added. This is done because in some cases (such as
* redeclarations) the declaration may not actually be added to the
* instruction stream.
*/
exec_list initializer_instructions;
/* Examine var name here since var may get deleted in the next call */
bool var_is_gl_id = is_gl_identifier(var->name);
bool is_redeclaration;
var = get_variable_being_redeclared(&var, decl->get_location(), state,
false /* allow_all_redeclarations */,
&is_redeclaration);
if (is_redeclaration) {
if (var_is_gl_id &&
var->data.how_declared == ir_var_declared_in_block) {
_mesa_glsl_error(&loc, state,
"`%s' has already been redeclared using "
"gl_PerVertex", var->name);
}
var->data.how_declared = ir_var_declared_normally;
}
if (decl->initializer != NULL) {
result = process_initializer(var,
decl, this->type,
&initializer_instructions, state);
} else {
validate_array_dimensions(var_type, state, &loc);
}
/* From page 23 (page 29 of the PDF) of the GLSL 1.10 spec:
*
* "It is an error to write to a const variable outside of
* its declaration, so they must be initialized when
* declared."
*/
if (this->type->qualifier.flags.q.constant && decl->initializer == NULL) {
_mesa_glsl_error(& loc, state,
"const declaration of `%s' must be initialized",
decl->identifier);
}
if (state->es_shader) {
const glsl_type *const t = var->type;
/* Skip the unsized array check for TCS/TES/GS inputs & TCS outputs.
*
* The GL_OES_tessellation_shader spec says about inputs:
*
* "Declaring an array size is optional. If no size is specified,
* it will be taken from the implementation-dependent maximum
* patch size (gl_MaxPatchVertices)."
*
* and about TCS outputs:
*
* "If no size is specified, it will be taken from output patch
* size declared in the shader."
*
* The GL_OES_geometry_shader spec says:
*
* "All geometry shader input unsized array declarations will be
* sized by an earlier input primitive layout qualifier, when
* present, as per the following table."
*/
const bool implicitly_sized =
(var->data.mode == ir_var_shader_in &&
state->stage >= MESA_SHADER_TESS_CTRL &&
state->stage <= MESA_SHADER_GEOMETRY) ||
(var->data.mode == ir_var_shader_out &&
state->stage == MESA_SHADER_TESS_CTRL);
if (t->is_unsized_array() && !implicitly_sized)
/* Section 10.17 of the GLSL ES 1.00 specification states that
* unsized array declarations have been removed from the language.
* Arrays that are sized using an initializer are still explicitly
* sized. However, GLSL ES 1.00 does not allow array
* initializers. That is only allowed in GLSL ES 3.00.
*
* Section 4.1.9 (Arrays) of the GLSL ES 3.00 spec says:
*
* "An array type can also be formed without specifying a size
* if the definition includes an initializer:
*
* float x[] = float[2] (1.0, 2.0); // declares an array of size 2
* float y[] = float[] (1.0, 2.0, 3.0); // declares an array of size 3
*
* float a[5];
* float b[] = a;"
*/
_mesa_glsl_error(& loc, state,
"unsized array declarations are not allowed in "
"GLSL ES");
}
/* Section 4.4.6.1 Atomic Counter Layout Qualifiers of the GLSL 4.60 spec:
*
* "It is a compile-time error to declare an unsized array of
* atomic_uint"
*/
if (var->type->is_unsized_array() &&
var->type->without_array()->base_type == GLSL_TYPE_ATOMIC_UINT) {
_mesa_glsl_error(& loc, state,
"Unsized array of atomic_uint is not allowed");
}
/* If the declaration is not a redeclaration, there are a few additional
* semantic checks that must be applied. In addition, variable that was
* created for the declaration should be added to the IR stream.
*/
if (!is_redeclaration) {
validate_identifier(decl->identifier, loc, state);
/* Add the variable to the symbol table. Note that the initializer's
* IR was already processed earlier (though it hasn't been emitted
* yet), without the variable in scope.
*
* This differs from most C-like languages, but it follows the GLSL
* specification. From page 28 (page 34 of the PDF) of the GLSL 1.50
* spec:
*
* "Within a declaration, the scope of a name starts immediately
* after the initializer if present or immediately after the name
* being declared if not."
*/
if (!state->symbols->add_variable(var)) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state, "name `%s' already taken in the "
"current scope", decl->identifier);
continue;
}
/* Push the variable declaration to the top. It means that all the
* variable declarations will appear in a funny last-to-first order,
* but otherwise we run into trouble if a function is prototyped, a
* global var is decled, then the function is defined with usage of
* the global var. See glslparsertest's CorrectModule.frag.
* However, do not insert declarations before default precision statements
* or type declarations.
*/
ir_instruction* before_node = (ir_instruction*)instructions->get_head();
while (before_node && (before_node->ir_type == ir_type_precision || before_node->ir_type == ir_type_typedecl))
before_node = (ir_instruction*)before_node->next;
if (before_node)
before_node->insert_before(var);
else
instructions->push_head(var);
}
instructions->append_list(&initializer_instructions);
}
/* Generally, variable declarations do not have r-values. However,
* one is used for the declaration in
*
* while (bool b = some_condition()) {
* ...
* }
*
* so we return the rvalue from the last seen declaration here.
*/
return result;
}
ir_rvalue *
ast_parameter_declarator::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
const struct glsl_type *type;
const char *name = NULL;
YYLTYPE loc = this->get_location();
type = this->type->glsl_type(& name, state);
if (type == NULL) {
if (name != NULL) {
_mesa_glsl_error(& loc, state,
"invalid type `%s' in declaration of `%s'",
name, this->identifier);
} else {
_mesa_glsl_error(& loc, state,
"invalid type in declaration of `%s'",
this->identifier);
}
type = glsl_type::error_type;
}
/* From page 62 (page 68 of the PDF) of the GLSL 1.50 spec:
*
* "Functions that accept no input arguments need not use void in the
* argument list because prototypes (or definitions) are required and
* therefore there is no ambiguity when an empty argument list "( )" is
* declared. The idiom "(void)" as a parameter list is provided for
* convenience."
*
* Placing this check here prevents a void parameter being set up
* for a function, which avoids tripping up checks for main taking
* parameters and lookups of an unnamed symbol.
*/
if (type->is_void()) {
if (this->identifier != NULL)
_mesa_glsl_error(& loc, state,
"named parameter cannot have type `void'");
is_void = true;
return NULL;
}
if (formal_parameter && (this->identifier == NULL)) {
_mesa_glsl_error(& loc, state, "formal parameter lacks a name");
return NULL;
}
/* This only handles "vec4 foo[..]". The earlier specifier->glsl_type(...)
* call already handled the "vec4[..] foo" case.
*/
type = process_array_type(&loc, type, this->array_specifier, state);
if (!type->is_error() && type->is_unsized_array()) {
_mesa_glsl_error(&loc, state, "arrays passed as parameters must have "
"a declared size");
type = glsl_type::error_type;
}
is_void = false;
ir_variable *var = new(ctx)
ir_variable(type, this->identifier, ir_var_function_in);
/* Apply any specified qualifiers to the parameter declaration. Note that
* for function parameters the default mode is 'in'.
*/
apply_type_qualifier_to_variable(& this->type->qualifier, var, state, & loc,
true);
/* From section 4.1.7 of the GLSL 4.40 spec:
*
* "Opaque variables cannot be treated as l-values; hence cannot
* be used as out or inout function parameters, nor can they be
* assigned into."
*
* From section 4.1.7 of the ARB_bindless_texture spec:
*
* "Samplers can be used as l-values, so can be assigned into and used
* as "out" and "inout" function parameters."
*
* From section 4.1.X of the ARB_bindless_texture spec:
*
* "Images can be used as l-values, so can be assigned into and used as
* "out" and "inout" function parameters."
*/
if ((var->data.mode == ir_var_function_inout || var->data.mode == ir_var_function_out)
&& (type->contains_atomic() ||
(!state->has_bindless() && type->contains_opaque()))) {
_mesa_glsl_error(&loc, state, "out and inout parameters cannot "
"contain %s variables",
state->has_bindless() ? "atomic" : "opaque");
type = glsl_type::error_type;
}
/* From page 39 (page 45 of the PDF) of the GLSL 1.10 spec:
*
* "When calling a function, expressions that do not evaluate to
* l-values cannot be passed to parameters declared as out or inout."
*
* From page 32 (page 38 of the PDF) of the GLSL 1.10 spec:
*
* "Other binary or unary expressions, non-dereferenced arrays,
* function names, swizzles with repeated fields, and constants
* cannot be l-values."
*
* So for GLSL 1.10, passing an array as an out or inout parameter is not
* allowed. This restriction is removed in GLSL 1.20, and in GLSL ES.
*/
if ((var->data.mode == ir_var_function_inout || var->data.mode == ir_var_function_out)
&& type->is_array()
&& !state->check_version(120, 100, &loc,
"arrays cannot be out or inout parameters")) {
type = glsl_type::error_type;
}
instructions->push_tail(var);
/* Parameter declarations do not have r-values.
*/
return NULL;
}
void
ast_parameter_declarator::parameters_to_hir(exec_list *ast_parameters,
bool formal,
exec_list *ir_parameters,
_mesa_glsl_parse_state *state)
{
ast_parameter_declarator *void_param = NULL;
unsigned count = 0;
foreach_list_typed (ast_parameter_declarator, param, link, ast_parameters) {
param->formal_parameter = formal;
param->hir(ir_parameters, state);
if (param->is_void)
void_param = param;
count++;
}
if ((void_param != NULL) && (count > 1)) {
YYLTYPE loc = void_param->get_location();
_mesa_glsl_error(& loc, state,
"`void' parameter must be only parameter");
}
}
void
emit_function(_mesa_glsl_parse_state *state, ir_function *f)
{
/* IR invariants disallow function declarations or definitions
* nested within other function definitions. But there is no
* requirement about the relative order of function declarations
* and definitions with respect to one another. So simply insert
* the new ir_function block at the end of the toplevel instruction
* list.
*/
state->toplevel_ir->push_tail(f);
}
ir_rvalue *
ast_function::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
ir_function *f = NULL;
ir_function_signature *sig = NULL;
exec_list hir_parameters;
YYLTYPE loc = this->get_location();
const char *const name = identifier;
/* New functions are always added to the top-level IR instruction stream,
* so this instruction list pointer is ignored. See also emit_function
* (called below).
*/
(void) instructions;
/* From page 21 (page 27 of the PDF) of the GLSL 1.20 spec,
*
* "Function declarations (prototypes) cannot occur inside of functions;
* they must be at global scope, or for the built-in functions, outside
* the global scope."
*
* From page 27 (page 33 of the PDF) of the GLSL ES 1.00.16 spec,
*
* "User defined functions may only be defined within the global scope."
*
* Note that this language does not appear in GLSL 1.10.
*/
if ((state->current_function != NULL) &&
state->is_version(120, 100)) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state,
"declaration of function `%s' not allowed within "
"function body", name);
}
validate_identifier(name, this->get_location(), state);
/* Convert the list of function parameters to HIR now so that they can be
* used below to compare this function's signature with previously seen
* signatures for functions with the same name.
*/
ast_parameter_declarator::parameters_to_hir(& this->parameters,
is_definition,
& hir_parameters, state);
const char *return_type_name;
const glsl_type *return_type =
this->return_type->glsl_type(& return_type_name, state);
if (!return_type) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state,
"function `%s' has undeclared return type `%s'",
name, return_type_name);
return_type = glsl_type::error_type;
}
/* ARB_shader_subroutine states:
* "Subroutine declarations cannot be prototyped. It is an error to prepend
* subroutine(...) to a function declaration."
*/
if (this->return_type->qualifier.subroutine_list && !is_definition) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state,
"function declaration `%s' cannot have subroutine prepended",
name);
}
/* From page 56 (page 62 of the PDF) of the GLSL 1.30 spec:
* "No qualifier is allowed on the return type of a function."
*/
if (this->return_type->has_qualifiers(state)) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"function `%s' return type has qualifiers", name);
}
/* Section 6.1 (Function Definitions) of the GLSL 1.20 spec says:
*
* "Arrays are allowed as arguments and as the return type. In both
* cases, the array must be explicitly sized."
*/
if (return_type->is_unsized_array()) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"function `%s' return type array must be explicitly "
"sized", name);
}
/* From Section 6.1 (Function Definitions) of the GLSL 1.00 spec:
*
* "Arrays are allowed as arguments, but not as the return type. [...]
* The return type can also be a structure if the structure does not
* contain an array."
*/
if (state->language_version == 100 && return_type->contains_array()) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"function `%s' return type contains an array", name);
}
/* From section 4.1.7 of the GLSL 4.40 spec:
*
* "[Opaque types] can only be declared as function parameters
* or uniform-qualified variables."
*
* The ARB_bindless_texture spec doesn't clearly state this, but as it says
* "Replace Section 4.1.7 (Samplers), p. 25" and, "Replace Section 4.1.X,
* (Images)", this should be allowed.
*/
if (return_type->contains_atomic() ||
(!state->has_bindless() && return_type->contains_opaque())) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state,
"function `%s' return type can't contain an %s type",
name, state->has_bindless() ? "atomic" : "opaque");
}
/**/
if (return_type->is_subroutine()) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state,
"function `%s' return type can't be a subroutine type",
name);
}
/* Get the precision for the return type */
unsigned return_precision;
if (state->es_shader) {
YYLTYPE loc = this->get_location();
return_precision =
select_gles_precision(this->return_type->qualifier.precision,
return_type,
state,
&loc);
} else {
return_precision = GLSL_PRECISION_NONE;
}
/* Create an ir_function if one doesn't already exist. */
f = state->symbols->get_function(name);
if (f == NULL) {
f = new(ctx) ir_function(name);
if (!this->return_type->qualifier.is_subroutine_decl()) {
if (!state->symbols->add_function(f)) {
/* This function name shadows a non-function use of the same name. */
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state, "function name `%s' conflicts with "
"non-function", name);
return NULL;
}
}
emit_function(state, f);
}
/* From GLSL ES 3.0 spec, chapter 6.1 "Function Definitions", page 71:
*
* "A shader cannot redefine or overload built-in functions."
*
* While in GLSL ES 1.0 specification, chapter 8 "Built-in Functions":
*
* "User code can overload the built-in functions but cannot redefine
* them."
*/
if (state->es_shader) {
/* Local shader has no exact candidates; check the built-ins. */
if (state->language_version >= 300 &&
_mesa_glsl_has_builtin_function(state, name)) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"A shader cannot redefine or overload built-in "
"function `%s' in GLSL ES 3.00", name);
return NULL;
}
if (state->language_version == 100) {
ir_function_signature *sig =
_mesa_glsl_find_builtin_function(state, name, &hir_parameters);
if (sig && sig->is_builtin()) {
_mesa_glsl_error(& loc, state,
"A shader cannot redefine built-in "
"function `%s' in GLSL ES 1.00", name);
}
}
}
/* Verify that this function's signature either doesn't match a previously
* seen signature for a function with the same name, or, if a match is found,
* that the previously seen signature does not have an associated definition.
*/
if (state->es_shader || f->has_user_signature()) {
sig = f->exact_matching_signature(state, &hir_parameters);
if (sig != NULL) {
const char *badvar = sig->qualifiers_match(&hir_parameters);
if (badvar != NULL) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state, "function `%s' parameter `%s' "
"qualifiers don't match prototype", name, badvar);
}
if (sig->return_type != return_type) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state, "function `%s' return type doesn't "
"match prototype", name);
}
if (sig->return_precision != return_precision) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state, "function `%s' return type precision "
"doesn't match prototype", name);
}
if (sig->is_defined) {
if (is_definition) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state, "function `%s' redefined", name);
} else {
/* We just encountered a prototype that exactly matches a
* function that's already been defined. This is redundant,
* and we should ignore it.
*/
return NULL;
}
} else if (state->language_version == 100 && !is_definition) {
/* From the GLSL 1.00 spec, section 4.2.7:
*
* "A particular variable, structure or function declaration
* may occur at most once within a scope with the exception
* that a single function prototype plus the corresponding
* function definition are allowed."
*/
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state, "function `%s' redeclared", name);
}
}
}
/* Verify the return type of main() */
if (strcmp(name, "main") == 0) {
if (! return_type->is_void()) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state, "main() must return void");
}
if (!hir_parameters.is_empty()) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state, "main() must not take any parameters");
}
}
/* Finish storing the information about this new function in its signature.
*/
if (sig == NULL) {
sig = new(ctx) ir_function_signature(return_type);
sig->return_precision = return_precision;
f->add_signature(sig);
}
sig->replace_parameters(&hir_parameters);
signature = sig;
if (this->return_type->qualifier.subroutine_list) {
int idx;
if (this->return_type->qualifier.flags.q.explicit_index) {
unsigned qual_index;
if (process_qualifier_constant(state, &loc, "index",
this->return_type->qualifier.index,
&qual_index)) {
if (!state->has_explicit_uniform_location()) {
_mesa_glsl_error(&loc, state, "subroutine index requires "
"GL_ARB_explicit_uniform_location or "
"GLSL 4.30");
} else if (qual_index >= MAX_SUBROUTINES) {
_mesa_glsl_error(&loc, state,
"invalid subroutine index (%d) index must "
"be a number between 0 and "
"GL_MAX_SUBROUTINES - 1 (%d)", qual_index,
MAX_SUBROUTINES - 1);
} else {
f->subroutine_index = qual_index;
}
}
}
f->num_subroutine_types = this->return_type->qualifier.subroutine_list->declarations.length();
f->subroutine_types = ralloc_array(state, const struct glsl_type *,
f->num_subroutine_types);
idx = 0;
foreach_list_typed(ast_declaration, decl, link, &this->return_type->qualifier.subroutine_list->declarations) {
const struct glsl_type *type;
/* the subroutine type must be already declared */
type = state->symbols->get_type(decl->identifier);
if (!type) {
_mesa_glsl_error(& loc, state, "unknown type '%s' in subroutine function definition", decl->identifier);
}
for (int i = 0; i < state->num_subroutine_types; i++) {
ir_function *fn = state->subroutine_types[i];
ir_function_signature *tsig = NULL;
if (strcmp(fn->name, decl->identifier))
continue;
tsig = fn->matching_signature(state, &sig->parameters,
false);
if (!tsig) {
_mesa_glsl_error(& loc, state, "subroutine type mismatch '%s' - signatures do not match\n", decl->identifier);
} else {
if (tsig->return_type != sig->return_type) {
_mesa_glsl_error(& loc, state, "subroutine type mismatch '%s' - return types do not match\n", decl->identifier);
}
}
}
f->subroutine_types[idx++] = type;
}
state->subroutines = (ir_function **)reralloc(state, state->subroutines,
ir_function *,
state->num_subroutines + 1);
state->subroutines[state->num_subroutines] = f;
state->num_subroutines++;
}
if (this->return_type->qualifier.is_subroutine_decl()) {
if (!state->symbols->add_type(this->identifier, glsl_type::get_subroutine_instance(this->identifier))) {
_mesa_glsl_error(& loc, state, "type '%s' previously defined", this->identifier);
return NULL;
}
state->subroutine_types = (ir_function **)reralloc(state, state->subroutine_types,
ir_function *,
state->num_subroutine_types + 1);
state->subroutine_types[state->num_subroutine_types] = f;
state->num_subroutine_types++;
f->is_subroutine = true;
}
/* Function declarations (prototypes) do not have r-values.
*/
return NULL;
}
ir_rvalue *
ast_function_definition::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
prototype->is_definition = true;
prototype->hir(instructions, state);
ir_function_signature *signature = prototype->signature;
if (signature == NULL)
return NULL;
assert(state->current_function == NULL);
state->current_function = signature;
state->found_return = false;
state->found_begin_interlock = false;
state->found_end_interlock = false;
/* Duplicate parameters declared in the prototype as concrete variables.
* Add these to the symbol table.
*/
state->symbols->push_scope();
foreach_in_list(ir_variable, var, &signature->parameters) {
assert(var->as_variable() != NULL);
/* The only way a parameter would "exist" is if two parameters have
* the same name.
*/
if (state->symbols->name_declared_this_scope(var->name)) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state, "parameter `%s' redeclared", var->name);
} else {
state->symbols->add_variable(var);
}
}
/* Convert the body of the function to HIR. */
this->body->hir(&signature->body, state);
signature->is_defined = true;
state->symbols->pop_scope();
assert(state->current_function == signature);
state->current_function = NULL;
if (!signature->return_type->is_void() && !state->found_return) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state, "function `%s' has non-void return type "
"%s, but no return statement",
signature->function_name(),
signature->return_type->name);
}
/* Function definitions do not have r-values.
*/
return NULL;
}
ir_rvalue *
ast_jump_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
switch (mode) {
case ast_return: {
ir_return *inst;
assert(state->current_function);
if (opt_return_value) {
ir_rvalue *ret = opt_return_value->hir(instructions, state);
/* The value of the return type can be NULL if the shader says
* 'return foo();' and foo() is a function that returns void.
*
* NOTE: The GLSL spec doesn't say that this is an error. The type
* of the return value is void. If the return type of the function is
* also void, then this should compile without error. Seriously.
*/
const glsl_type *const ret_type =
(ret == NULL) ? glsl_type::void_type : ret->type;
/* Implicit conversions are not allowed for return values prior to
* ARB_shading_language_420pack.
*/
if (state->current_function->return_type != ret_type) {
YYLTYPE loc = this->get_location();
if (state->has_420pack()) {
if (!apply_implicit_conversion(state->current_function->return_type,
ret, state)
|| (ret->type != state->current_function->return_type)) {
_mesa_glsl_error(& loc, state,
"could not implicitly convert return value "
"to %s, in function `%s'",
state->current_function->return_type->name,
state->current_function->function_name());
}
} else {
_mesa_glsl_error(& loc, state,
"`return' with wrong type %s, in function `%s' "
"returning %s",
ret_type->name,
state->current_function->function_name(),
state->current_function->return_type->name);
}
} else if (state->current_function->return_type->base_type ==
GLSL_TYPE_VOID) {
YYLTYPE loc = this->get_location();
/* The ARB_shading_language_420pack, GLSL ES 3.0, and GLSL 4.20
* specs add a clarification:
*
* "A void function can only use return without a return argument, even if
* the return argument has void type. Return statements only accept values:
*
* void func1() { }
* void func2() { return func1(); } // illegal return statement"
*/
_mesa_glsl_error(& loc, state,
"void functions can only use `return' without a "
"return argument");
}
inst = new(ctx) ir_return(ret);
} else {
if (state->current_function->return_type->base_type !=
GLSL_TYPE_VOID) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"`return' with no value, in function %s returning "
"non-void",
state->current_function->function_name());
}
inst = new(ctx) ir_return;
}
state->found_return = true;
instructions->push_tail(inst);
break;
}
case ast_discard:
if (state->stage != MESA_SHADER_FRAGMENT) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"`discard' may only appear in a fragment shader");
}
instructions->push_tail(new(ctx) ir_discard);
break;
case ast_break:
case ast_continue:
if (mode == ast_continue &&
state->loop_nesting_ast == NULL) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state, "continue may only appear in a loop");
} else if (mode == ast_break &&
state->loop_nesting_ast == NULL &&
state->switch_state.switch_nesting_ast == NULL) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"break may only appear in a loop or a switch");
} else {
/* For a loop, inline the for loop expression again, since we don't
* know where near the end of the loop body the normal copy of it is
* going to be placed. Same goes for the condition for a do-while
* loop.
*/
if (state->loop_nesting_ast != NULL &&
mode == ast_continue) {
if (state->loop_nesting_ast->rest_expression) {
state->loop_nesting_ast->rest_expression->hir(instructions,
state);
}
if (state->loop_nesting_ast->mode ==
ast_iteration_statement::ast_do_while) {
state->loop_nesting_ast->condition_to_hir(instructions, state);
}
}
if (state->switch_state.is_switch_innermost &&
mode == ast_break) {
/* Force break out of switch by setting is_break switch state.
*/
ir_variable *const is_break_var = state->switch_state.is_break_var;
ir_dereference_variable *const deref_is_break_var =
new(ctx) ir_dereference_variable(is_break_var);
ir_constant *const true_val = new(ctx) ir_constant(true);
ir_assignment *const set_break_var =
new(ctx) ir_assignment(deref_is_break_var, true_val);
instructions->push_tail(set_break_var);
} else {
ir_loop_jump *const jump =
new(ctx) ir_loop_jump((mode == ast_break)
? ir_loop_jump::jump_break
: ir_loop_jump::jump_continue);
instructions->push_tail(jump);
}
}
break;
}
/* Jump instructions do not have r-values.
*/
return NULL;
}
ir_rvalue *
ast_demote_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
if (state->stage != MESA_SHADER_FRAGMENT) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"`demote' may only appear in a fragment shader");
}
instructions->push_tail(new(ctx) ir_demote);
return NULL;
}
ir_rvalue *
ast_selection_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
ir_rvalue *const condition = this->condition->hir(instructions, state);
/* From page 66 (page 72 of the PDF) of the GLSL 1.50 spec:
*
* "Any expression whose type evaluates to a Boolean can be used as the
* conditional expression bool-expression. Vector types are not accepted
* as the expression to if."
*
* The checks are separated so that higher quality diagnostics can be
* generated for cases where both rules are violated.
*/
if (!condition->type->is_boolean() || !condition->type->is_scalar()) {
YYLTYPE loc = this->condition->get_location();
_mesa_glsl_error(& loc, state, "if-statement condition must be scalar "
"boolean");
}
ir_if *const stmt = new(ctx) ir_if(condition);
if (then_statement != NULL) {
state->symbols->push_scope();
then_statement->hir(& stmt->then_instructions, state);
state->symbols->pop_scope();
}
if (else_statement != NULL) {
state->symbols->push_scope();
else_statement->hir(& stmt->else_instructions, state);
state->symbols->pop_scope();
}
instructions->push_tail(stmt);
/* if-statements do not have r-values.
*/
return NULL;
}
struct case_label {
/** Value of the case label. */
unsigned value;
/** Does this label occur after the default? */
bool after_default;
/**
* AST for the case label.
*
* This is only used to generate error messages for duplicate labels.
*/
ast_expression *ast;
};
/* Used for detection of duplicate case values, compare
* given contents directly.
*/
static bool
compare_case_value(const void *a, const void *b)
{
return ((struct case_label *) a)->value == ((struct case_label *) b)->value;
}
/* Used for detection of duplicate case values, just
* returns key contents as is.
*/
static unsigned
key_contents(const void *key)
{
return ((struct case_label *) key)->value;
}
ir_rvalue *
ast_switch_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
ir_rvalue *const test_expression =
this->test_expression->hir(instructions, state);
/* From page 66 (page 55 of the PDF) of the GLSL 1.50 spec:
*
* "The type of init-expression in a switch statement must be a
* scalar integer."
*/
if (!test_expression->type->is_scalar() ||
!test_expression->type->is_integer_32()) {
YYLTYPE loc = this->test_expression->get_location();
_mesa_glsl_error(& loc,
state,
"switch-statement expression must be scalar "
"integer");
return NULL;
}
/* Track the switch-statement nesting in a stack-like manner.
*/
struct glsl_switch_state saved = state->switch_state;
state->switch_state.is_switch_innermost = true;
state->switch_state.switch_nesting_ast = this;
state->switch_state.labels_ht =
_mesa_hash_table_create(NULL, key_contents,
compare_case_value);
state->switch_state.previous_default = NULL;
/* Initalize is_fallthru state to false.
*/
ir_rvalue *const is_fallthru_val = new (ctx) ir_constant(false);
state->switch_state.is_fallthru_var =
new(ctx) ir_variable(glsl_type::bool_type,
"switch_is_fallthru_tmp",
ir_var_temporary);
instructions->push_tail(state->switch_state.is_fallthru_var);
ir_dereference_variable *deref_is_fallthru_var =
new(ctx) ir_dereference_variable(state->switch_state.is_fallthru_var);
instructions->push_tail(new(ctx) ir_assignment(deref_is_fallthru_var,
is_fallthru_val));
/* Initialize is_break state to false.
*/
ir_rvalue *const is_break_val = new (ctx) ir_constant(false);
state->switch_state.is_break_var =
new(ctx) ir_variable(glsl_type::bool_type,
"switch_is_break_tmp",
ir_var_temporary);
instructions->push_tail(state->switch_state.is_break_var);
ir_dereference_variable *deref_is_break_var =
new(ctx) ir_dereference_variable(state->switch_state.is_break_var);
instructions->push_tail(new(ctx) ir_assignment(deref_is_break_var,
is_break_val));
state->switch_state.run_default =
new(ctx) ir_variable(glsl_type::bool_type,
"run_default_tmp",
ir_var_temporary);
instructions->push_tail(state->switch_state.run_default);
/* Cache test expression.
*/
test_to_hir(instructions, state);
/* Emit code for body of switch stmt.
*/
body->hir(instructions, state);
_mesa_hash_table_destroy(state->switch_state.labels_ht, NULL);
state->switch_state = saved;
/* Switch statements do not have r-values. */
return NULL;
}
void
ast_switch_statement::test_to_hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
/* set to true to avoid a duplicate "use of uninitialized variable" warning
* on the switch test case. The first one would be already raised when
* getting the test_expression at ast_switch_statement::hir
*/
test_expression->set_is_lhs(true);
/* Cache value of test expression. */
ir_rvalue *const test_val = test_expression->hir(instructions, state);
state->switch_state.test_var = new(ctx) ir_variable(test_val->type,
"switch_test_tmp",
ir_var_temporary);
ir_dereference_variable *deref_test_var =
new(ctx) ir_dereference_variable(state->switch_state.test_var);
instructions->push_tail(state->switch_state.test_var);
instructions->push_tail(new(ctx) ir_assignment(deref_test_var, test_val));
}
ir_rvalue *
ast_switch_body::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
if (stmts != NULL)
stmts->hir(instructions, state);
/* Switch bodies do not have r-values. */
return NULL;
}
ir_rvalue *
ast_case_statement_list::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
exec_list default_case, after_default, tmp;
foreach_list_typed (ast_case_statement, case_stmt, link, & this->cases) {
case_stmt->hir(&tmp, state);
/* Default case. */
if (state->switch_state.previous_default && default_case.is_empty()) {
default_case.append_list(&tmp);
continue;
}
/* If default case found, append 'after_default' list. */
if (!default_case.is_empty())
after_default.append_list(&tmp);
else
instructions->append_list(&tmp);
}
/* Handle the default case. This is done here because default might not be
* the last case. We need to add checks against following cases first to see
* if default should be chosen or not.
*/
if (!default_case.is_empty()) {
ir_factory body(instructions, state);
ir_expression *cmp = NULL;
hash_table_foreach(state->switch_state.labels_ht, entry) {
const struct case_label *const l = (struct case_label *) entry->data;
/* If the switch init-value is the value of one of the labels that
* occurs after the default case, disable execution of the default
* case.
*/
if (l->after_default) {
ir_constant *const cnst =
state->switch_state.test_var->type->base_type == GLSL_TYPE_UINT
? body.constant(unsigned(l->value))
: body.constant(int(l->value));
cmp = cmp == NULL
? equal(cnst, state->switch_state.test_var)
: logic_or(cmp, equal(cnst, state->switch_state.test_var));
}
}
if (cmp != NULL)
body.emit(assign(state->switch_state.run_default, logic_not(cmp)));
else
body.emit(assign(state->switch_state.run_default, body.constant(true)));
/* Append default case and all cases after it. */
instructions->append_list(&default_case);
instructions->append_list(&after_default);
}
/* Case statements do not have r-values. */
return NULL;
}
ir_rvalue *
ast_case_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
labels->hir(instructions, state);
/* Conditionally set fallthru state based on break state. */
ir_factory reset_fallthru(instructions, state);
reset_fallthru.emit(assign(state->switch_state.is_fallthru_var,
logic_and(state->switch_state.is_fallthru_var,
logic_not(state->switch_state.is_break_var))));
/* Guard case statements depending on fallthru state. */
ir_dereference_variable *const deref_fallthru_guard =
new(state) ir_dereference_variable(state->switch_state.is_fallthru_var);
ir_if *const test_fallthru = new(state) ir_if(deref_fallthru_guard);
foreach_list_typed (ast_node, stmt, link, & this->stmts)
stmt->hir(& test_fallthru->then_instructions, state);
instructions->push_tail(test_fallthru);
/* Case statements do not have r-values. */
return NULL;
}
ir_rvalue *
ast_case_label_list::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
foreach_list_typed (ast_case_label, label, link, & this->labels)
label->hir(instructions, state);
/* Case labels do not have r-values. */
return NULL;
}
ir_rvalue *
ast_case_label::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
ir_factory body(instructions, state);
ir_variable *const fallthru_var = state->switch_state.is_fallthru_var;
/* If not default case, ... */
if (this->test_value != NULL) {
/* Conditionally set fallthru state based on
* comparison of cached test expression value to case label.
*/
ir_rvalue *const label_rval = this->test_value->hir(instructions, state);
ir_constant *label_const =
label_rval->constant_expression_value(body.mem_ctx);
if (!label_const) {
YYLTYPE loc = this->test_value->get_location();
_mesa_glsl_error(& loc, state,
"switch statement case label must be a "
"constant expression");
/* Stuff a dummy value in to allow processing to continue. */
label_const = body.constant(0);
} else {
hash_entry *entry =
_mesa_hash_table_search(state->switch_state.labels_ht,
&label_const->value.u[0]);
if (entry) {
const struct case_label *const l =
(struct case_label *) entry->data;
const ast_expression *const previous_label = l->ast;
YYLTYPE loc = this->test_value->get_location();
_mesa_glsl_error(& loc, state, "duplicate case value");
loc = previous_label->get_location();
_mesa_glsl_error(& loc, state, "this is the previous case label");
} else {
struct case_label *l = ralloc(state->switch_state.labels_ht,
struct case_label);
l->value = label_const->value.u[0];
l->after_default = state->switch_state.previous_default != NULL;
l->ast = this->test_value;
_mesa_hash_table_insert(state->switch_state.labels_ht,
&label_const->value.u[0],
l);
}
}
/* Create an r-value version of the ir_constant label here (after we may
* have created a fake one in error cases) that can be passed to
* apply_implicit_conversion below.
*/
ir_rvalue *label = label_const;
ir_rvalue *deref_test_var =
new(body.mem_ctx) ir_dereference_variable(state->switch_state.test_var);
/*
* From GLSL 4.40 specification section 6.2 ("Selection"):
*
* "The type of the init-expression value in a switch statement must
* be a scalar int or uint. The type of the constant-expression value
* in a case label also must be a scalar int or uint. When any pair
* of these values is tested for "equal value" and the types do not
* match, an implicit conversion will be done to convert the int to a
* uint (see section 4.1.10 “Implicit Conversions”) before the compare
* is done."
*/
if (label->type != state->switch_state.test_var->type) {
YYLTYPE loc = this->test_value->get_location();
const glsl_type *type_a = label->type;
const glsl_type *type_b = state->switch_state.test_var->type;
/* Check if int->uint implicit conversion is supported. */
bool integer_conversion_supported =
glsl_type::int_type->can_implicitly_convert_to(glsl_type::uint_type,
state);
if ((!type_a->is_integer_32() || !type_b->is_integer_32()) ||
!integer_conversion_supported) {
_mesa_glsl_error(&loc, state, "type mismatch with switch "
"init-expression and case label (%s != %s)",
type_a->name, type_b->name);
} else {
/* Conversion of the case label. */
if (type_a->base_type == GLSL_TYPE_INT) {
if (!apply_implicit_conversion(glsl_type::uint_type,
label, state))
_mesa_glsl_error(&loc, state, "implicit type conversion error");
} else {
/* Conversion of the init-expression value. */
if (!apply_implicit_conversion(glsl_type::uint_type,
deref_test_var, state))
_mesa_glsl_error(&loc, state, "implicit type conversion error");
}
}
/* If the implicit conversion was allowed, the types will already be
* the same. If the implicit conversion wasn't allowed, smash the
* type of the label anyway. This will prevent the expression
* constructor (below) from failing an assertion.
*/
label->type = deref_test_var->type;
}
body.emit(assign(fallthru_var,
logic_or(fallthru_var, equal(label, deref_test_var))));
} else { /* default case */
if (state->switch_state.previous_default) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(& loc, state,
"multiple default labels in one switch");
loc = state->switch_state.previous_default->get_location();
_mesa_glsl_error(& loc, state, "this is the first default label");
}
state->switch_state.previous_default = this;
/* Set fallthru condition on 'run_default' bool. */
body.emit(assign(fallthru_var,
logic_or(fallthru_var,
state->switch_state.run_default)));
}
/* Case statements do not have r-values. */
return NULL;
}
void
ast_iteration_statement::condition_to_hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
if (condition != NULL) {
ir_rvalue *const cond =
condition->hir(instructions, state);
if ((cond == NULL)
|| !cond->type->is_boolean() || !cond->type->is_scalar()) {
YYLTYPE loc = condition->get_location();
_mesa_glsl_error(& loc, state,
"loop condition must be scalar boolean");
} else {
/* As the first code in the loop body, generate a block that looks
* like 'if (!condition) break;' as the loop termination condition.
*/
ir_rvalue *const not_cond =
new(ctx) ir_expression(ir_unop_logic_not, cond);
ir_if *const if_stmt = new(ctx) ir_if(not_cond);
ir_jump *const break_stmt =
new(ctx) ir_loop_jump(ir_loop_jump::jump_break);
if_stmt->then_instructions.push_tail(break_stmt);
instructions->push_tail(if_stmt);
}
}
}
ir_rvalue *
ast_iteration_statement::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
void *ctx = state;
/* For-loops and while-loops start a new scope, but do-while loops do not.
*/
if (mode != ast_do_while)
state->symbols->push_scope();
if (init_statement != NULL)
init_statement->hir(instructions, state);
ir_loop *const stmt = new(ctx) ir_loop();
instructions->push_tail(stmt);
/* Track the current loop nesting. */
ast_iteration_statement *nesting_ast = state->loop_nesting_ast;
state->loop_nesting_ast = this;
/* Likewise, indicate that following code is closest to a loop,
* NOT closest to a switch.
*/
bool saved_is_switch_innermost = state->switch_state.is_switch_innermost;
state->switch_state.is_switch_innermost = false;
if (mode != ast_do_while)
condition_to_hir(&stmt->body_instructions, state);
if (body != NULL)
body->hir(& stmt->body_instructions, state);
if (rest_expression != NULL)
rest_expression->hir(& stmt->body_instructions, state);
if (mode == ast_do_while)
condition_to_hir(&stmt->body_instructions, state);
if (mode != ast_do_while)
state->symbols->pop_scope();
/* Restore previous nesting before returning. */
state->loop_nesting_ast = nesting_ast;
state->switch_state.is_switch_innermost = saved_is_switch_innermost;
/* Loops do not have r-values.
*/
return NULL;
}
/**
* Determine if the given type is valid for establishing a default precision
* qualifier.
*
* From GLSL ES 3.00 section 4.5.4 ("Default Precision Qualifiers"):
*
* "The precision statement
*
* precision precision-qualifier type;
*
* can be used to establish a default precision qualifier. The type field
* can be either int or float or any of the sampler types, and the
* precision-qualifier can be lowp, mediump, or highp."
*
* GLSL ES 1.00 has similar language. GLSL 1.30 doesn't allow precision
* qualifiers on sampler types, but this seems like an oversight (since the
* intention of including these in GLSL 1.30 is to allow compatibility with ES
* shaders). So we allow int, float, and all sampler types regardless of GLSL
* version.
*/
static bool
is_valid_default_precision_type(const struct glsl_type *const type)
{
if (type == NULL)
return false;
switch (type->base_type) {
case GLSL_TYPE_INT:
case GLSL_TYPE_FLOAT:
/* "int" and "float" are valid, but vectors and matrices are not. */
return type->vector_elements == 1 && type->matrix_columns == 1;
case GLSL_TYPE_SAMPLER:
case GLSL_TYPE_IMAGE:
case GLSL_TYPE_ATOMIC_UINT:
return true;
default:
return false;
}
}
ir_rvalue *
ast_type_specifier::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
if (this->default_precision == ast_precision_none && this->structure == NULL)
return NULL;
YYLTYPE loc = this->get_location();
/* If this is a precision statement, check that the type to which it is
* applied is either float or int.
*
* From section 4.5.3 of the GLSL 1.30 spec:
* "The precision statement
* precision precision-qualifier type;
* can be used to establish a default precision qualifier. The type
* field can be either int or float [...]. Any other types or
* qualifiers will result in an error.
*/
if (this->default_precision != ast_precision_none) {
if (!state->check_precision_qualifiers_allowed(&loc))
return NULL;
if (this->structure != NULL) {
_mesa_glsl_error(&loc, state,
"precision qualifiers do not apply to structures");
return NULL;
}
if (this->array_specifier != NULL) {
_mesa_glsl_error(&loc, state,
"default precision statements do not apply to "
"arrays");
return NULL;
}
const struct glsl_type *const type =
state->symbols->get_type(this->type_name);
if (!is_valid_default_precision_type(type)) {
_mesa_glsl_error(&loc, state,
"default precision statements apply only to "
"float, int, and opaque types");
return NULL;
}
if (state->es_shader) {
/* Section 4.5.3 (Default Precision Qualifiers) of the GLSL ES 1.00
* spec says:
*
* "Non-precision qualified declarations will use the precision
* qualifier specified in the most recent precision statement
* that is still in scope. The precision statement has the same
* scoping rules as variable declarations. If it is declared
* inside a compound statement, its effect stops at the end of
* the innermost statement it was declared in. Precision
* statements in nested scopes override precision statements in
* outer scopes. Multiple precision statements for the same basic
* type can appear inside the same scope, with later statements
* overriding earlier statements within that scope."
*
* Default precision specifications follow the same scope rules as
* variables. So, we can track the state of the default precision
* qualifiers in the symbol table, and the rules will just work. This
* is a slight abuse of the symbol table, but it has the semantics
* that we want.
*/
state->symbols->add_default_precision_qualifier(this->type_name,
this->default_precision);
}
{
void *ctx = state;
const char* precision_type = NULL;
switch (this->default_precision) {
case GLSL_PRECISION_HIGH:
precision_type = "highp";
break;
case GLSL_PRECISION_MEDIUM:
precision_type = "mediump";
break;
case GLSL_PRECISION_LOW:
precision_type = "lowp";
break;
case GLSL_PRECISION_NONE:
precision_type = "";
break;
}
char* precision_statement = ralloc_asprintf(ctx, "precision %s %s", precision_type, this->type_name);
ir_precision_statement *const stmt = new(ctx) ir_precision_statement(precision_statement);
instructions->push_head(stmt);
}
return NULL;
}
/* _mesa_ast_set_aggregate_type() sets the <structure> field so that
* process_record_constructor() can do type-checking on C-style initializer
* expressions of structs, but ast_struct_specifier should only be translated
* to HIR if it is declaring the type of a structure.
*
* The ->is_declaration field is false for initializers of variables
* declared separately from the struct's type definition.
*
* struct S { ... }; (is_declaration = true)
* struct T { ... } t = { ... }; (is_declaration = true)
* S s = { ... }; (is_declaration = false)
*/
if (this->structure != NULL && this->structure->is_declaration)
return this->structure->hir(instructions, state);
return NULL;
}
/**
* Process a structure or interface block tree into an array of structure fields
*
* After parsing, where there are some syntax differnces, structures and
* interface blocks are almost identical. They are similar enough that the
* AST for each can be processed the same way into a set of
* \c glsl_struct_field to describe the members.
*
* If we're processing an interface block, var_mode should be the type of the
* interface block (ir_var_shader_in, ir_var_shader_out, ir_var_uniform or
* ir_var_shader_storage). If we're processing a structure, var_mode should be
* ir_var_auto.
*
* \return
* The number of fields processed. A pointer to the array structure fields is
* stored in \c *fields_ret.
*/
static unsigned
ast_process_struct_or_iface_block_members(exec_list *instructions,
struct _mesa_glsl_parse_state *state,
exec_list *declarations,
glsl_struct_field **fields_ret,
bool is_interface,
enum glsl_matrix_layout matrix_layout,
bool allow_reserved_names,
ir_variable_mode var_mode,
ast_type_qualifier *layout,
unsigned block_stream,
unsigned block_xfb_buffer,
unsigned block_xfb_offset,
unsigned expl_location,
unsigned expl_align)
{
unsigned decl_count = 0;
unsigned next_offset = 0;
/* Make an initial pass over the list of fields to determine how
* many there are. Each element in this list is an ast_declarator_list.
* This means that we actually need to count the number of elements in the
* 'declarations' list in each of the elements.
*/
foreach_list_typed (ast_declarator_list, decl_list, link, declarations) {
decl_count += decl_list->declarations.length();
}
/* Allocate storage for the fields and process the field
* declarations. As the declarations are processed, try to also convert
* the types to HIR. This ensures that structure definitions embedded in
* other structure definitions or in interface blocks are processed.
*/
glsl_struct_field *const fields = rzalloc_array(state, glsl_struct_field,
decl_count);
bool first_member = true;
bool first_member_has_explicit_location = false;
unsigned i = 0;
foreach_list_typed (ast_declarator_list, decl_list, link, declarations) {
const char *type_name;
YYLTYPE loc = decl_list->get_location();
decl_list->type->specifier->hir(instructions, state);
/* Section 4.1.8 (Structures) of the GLSL 1.10 spec says:
*
* "Anonymous structures are not supported; so embedded structures
* must have a declarator. A name given to an embedded struct is
* scoped at the same level as the struct it is embedded in."
*
* The same section of the GLSL 1.20 spec says:
*
* "Anonymous structures are not supported. Embedded structures are
* not supported."
*
* The GLSL ES 1.00 and 3.00 specs have similar langauge. So, we allow
* embedded structures in 1.10 only.
*/
if (state->language_version != 110 &&
decl_list->type->specifier->structure != NULL)
_mesa_glsl_error(&loc, state,
"embedded structure declarations are not allowed");
const glsl_type *decl_type =
decl_list->type->glsl_type(& type_name, state);
const struct ast_type_qualifier *const qual =
&decl_list->type->qualifier;
/* From section 4.3.9 of the GLSL 4.40 spec:
*
* "[In interface blocks] opaque types are not allowed."
*
* It should be impossible for decl_type to be NULL here. Cases that
* might naturally lead to decl_type being NULL, especially for the
* is_interface case, will have resulted in compilation having
* already halted due to a syntax error.
*/
assert(decl_type);
if (is_interface) {
/* From section 4.3.7 of the ARB_bindless_texture spec:
*
* "(remove the following bullet from the last list on p. 39,
* thereby permitting sampler types in interface blocks; image
* types are also permitted in blocks by this extension)"
*
* * sampler types are not allowed
*/
if (decl_type->contains_atomic() ||
(!state->has_bindless() && decl_type->contains_opaque())) {
_mesa_glsl_error(&loc, state, "uniform/buffer in non-default "
"interface block contains %s variable",
state->has_bindless() ? "atomic" : "opaque");
}
} else {
if (decl_type->contains_atomic()) {
/* From section 4.1.7.3 of the GLSL 4.40 spec:
*
* "Members of structures cannot be declared as atomic counter
* types."
*/
_mesa_glsl_error(&loc, state, "atomic counter in structure");
}
if (!state->has_bindless() && decl_type->contains_image()) {
/* FINISHME: Same problem as with atomic counters.
* FINISHME: Request clarification from Khronos and add
* FINISHME: spec quotation here.
*/
_mesa_glsl_error(&loc, state, "image in structure");
}
}
if (qual->flags.q.explicit_binding) {
_mesa_glsl_error(&loc, state,
"binding layout qualifier cannot be applied "
"to struct or interface block members");
}
if (is_interface) {
if (!first_member) {
if (!layout->flags.q.explicit_location &&
((first_member_has_explicit_location &&
!qual->flags.q.explicit_location) ||
(!first_member_has_explicit_location &&
qual->flags.q.explicit_location))) {
_mesa_glsl_error(&loc, state,
"when block-level location layout qualifier "
"is not supplied either all members must "
"have a location layout qualifier or all "
"members must not have a location layout "
"qualifier");
}
} else {
first_member = false;
first_member_has_explicit_location =
qual->flags.q.explicit_location;
}
}
if (qual->flags.q.std140 ||
qual->flags.q.std430 ||
qual->flags.q.packed ||
qual->flags.q.shared) {
_mesa_glsl_error(&loc, state,
"uniform/shader storage block layout qualifiers "
"std140, std430, packed, and shared can only be "
"applied to uniform/shader storage blocks, not "
"members");
}
if (qual->flags.q.constant) {
_mesa_glsl_error(&loc, state,
"const storage qualifier cannot be applied "
"to struct or interface block members");
}
validate_memory_qualifier_for_type(state, &loc, qual, decl_type);
validate_image_format_qualifier_for_type(state, &loc, qual, decl_type);
/* From Section 4.4.2.3 (Geometry Outputs) of the GLSL 4.50 spec:
*
* "A block member may be declared with a stream identifier, but
* the specified stream must match the stream associated with the
* containing block."
*/
if (qual->flags.q.explicit_stream) {
unsigned qual_stream;
if (process_qualifier_constant(state, &loc, "stream",
qual->stream, &qual_stream) &&
qual_stream != block_stream) {
_mesa_glsl_error(&loc, state, "stream layout qualifier on "
"interface block member does not match "
"the interface block (%u vs %u)", qual_stream,
block_stream);
}
}
int xfb_buffer;
unsigned explicit_xfb_buffer = 0;
if (qual->flags.q.explicit_xfb_buffer) {
unsigned qual_xfb_buffer;
if (process_qualifier_constant(state, &loc, "xfb_buffer",
qual->xfb_buffer, &qual_xfb_buffer)) {
explicit_xfb_buffer = 1;
if (qual_xfb_buffer != block_xfb_buffer)
_mesa_glsl_error(&loc, state, "xfb_buffer layout qualifier on "
"interface block member does not match "
"the interface block (%u vs %u)",
qual_xfb_buffer, block_xfb_buffer);
}
xfb_buffer = (int) qual_xfb_buffer;
} else {
if (layout)
explicit_xfb_buffer = layout->flags.q.explicit_xfb_buffer;
xfb_buffer = (int) block_xfb_buffer;
}
int xfb_stride = -1;
if (qual->flags.q.explicit_xfb_stride) {
unsigned qual_xfb_stride;
if (process_qualifier_constant(state, &loc, "xfb_stride",
qual->xfb_stride, &qual_xfb_stride)) {
xfb_stride = (int) qual_xfb_stride;
}
}
if (qual->flags.q.uniform && qual->has_interpolation()) {
_mesa_glsl_error(&loc, state,
"interpolation qualifiers cannot be used "
"with uniform interface blocks");
}
if ((qual->flags.q.uniform || !is_interface) &&
qual->has_auxiliary_storage()) {
_mesa_glsl_error(&loc, state,
"auxiliary storage qualifiers cannot be used "
"in uniform blocks or structures.");
}
if (qual->flags.q.row_major || qual->flags.q.column_major) {
if (!qual->flags.q.uniform && !qual->flags.q.buffer) {
_mesa_glsl_error(&loc, state,
"row_major and column_major can only be "
"applied to interface blocks");
} else
validate_matrix_layout_for_type(state, &loc, decl_type, NULL);
}
foreach_list_typed (ast_declaration, decl, link,
&decl_list->declarations) {
YYLTYPE loc = decl->get_location();
if (!allow_reserved_names)
validate_identifier(decl->identifier, loc, state);
const struct glsl_type *field_type =
process_array_type(&loc, decl_type, decl->array_specifier, state);
validate_array_dimensions(field_type, state, &loc);
fields[i].type = field_type;
fields[i].name = decl->identifier;
fields[i].interpolation =
interpret_interpolation_qualifier(qual, field_type,
var_mode, state, &loc);
fields[i].centroid = qual->flags.q.centroid ? 1 : 0;
fields[i].sample = qual->flags.q.sample ? 1 : 0;
fields[i].patch = qual->flags.q.patch ? 1 : 0;
fields[i].offset = -1;
fields[i].explicit_xfb_buffer = explicit_xfb_buffer;
fields[i].xfb_buffer = xfb_buffer;
fields[i].xfb_stride = xfb_stride;
if (qual->flags.q.explicit_location) {
unsigned qual_location;
if (process_qualifier_constant(state, &loc, "location",
qual->location, &qual_location)) {
fields[i].location = qual_location +
(fields[i].patch ? VARYING_SLOT_PATCH0 : VARYING_SLOT_VAR0);
expl_location = fields[i].location +
fields[i].type->count_attribute_slots(false);
}
} else {
if (layout && layout->flags.q.explicit_location) {
fields[i].location = expl_location;
expl_location += fields[i].type->count_attribute_slots(false);
} else {
fields[i].location = -1;
}
}
/* Offset can only be used with std430 and std140 layouts an initial
* value of 0 is used for error detection.
*/
unsigned align = 0;
unsigned size = 0;
if (layout) {
bool row_major;
if (qual->flags.q.row_major ||
matrix_layout == GLSL_MATRIX_LAYOUT_ROW_MAJOR) {
row_major = true;
} else {
row_major = false;
}
if(layout->flags.q.std140) {
align = field_type->std140_base_alignment(row_major);
size = field_type->std140_size(row_major);
} else if (layout->flags.q.std430) {
align = field_type->std430_base_alignment(row_major);
size = field_type->std430_size(row_major);
}
}
if (qual->flags.q.explicit_offset) {
unsigned qual_offset;
if (process_qualifier_constant(state, &loc, "offset",
qual->offset, &qual_offset)) {
if (align != 0 && size != 0) {
if (next_offset > qual_offset)
_mesa_glsl_error(&loc, state, "layout qualifier "
"offset overlaps previous member");
if (qual_offset % align) {
_mesa_glsl_error(&loc, state, "layout qualifier offset "
"must be a multiple of the base "
"alignment of %s", field_type->name);
}
fields[i].offset = qual_offset;
next_offset = qual_offset + size;
} else {
_mesa_glsl_error(&loc, state, "offset can only be used "
"with std430 and std140 layouts");
}
}
}
if (qual->flags.q.explicit_align || expl_align != 0) {
unsigned offset = fields[i].offset != -1 ? fields[i].offset :
next_offset;
if (align == 0 || size == 0) {
_mesa_glsl_error(&loc, state, "align can only be used with "
"std430 and std140 layouts");
} else if (qual->flags.q.explicit_align) {
unsigned member_align;
if (process_qualifier_constant(state, &loc, "align",
qual->align, &member_align)) {
if (member_align == 0 ||
member_align & (member_align - 1)) {
_mesa_glsl_error(&loc, state, "align layout qualifier "
"is not a power of 2");
} else {
fields[i].offset = glsl_align(offset, member_align);
next_offset = fields[i].offset + size;
}
}
} else {
fields[i].offset = glsl_align(offset, expl_align);
next_offset = fields[i].offset + size;
}
} else if (!qual->flags.q.explicit_offset) {
if (align != 0 && size != 0)
next_offset = glsl_align(next_offset, align) + size;
}
/* From the ARB_enhanced_layouts spec:
*
* "The given offset applies to the first component of the first
* member of the qualified entity. Then, within the qualified
* entity, subsequent components are each assigned, in order, to
* the next available offset aligned to a multiple of that
* component's size. Aggregate types are flattened down to the
* component level to get this sequence of components."
*/
if (qual->flags.q.explicit_xfb_offset) {
unsigned xfb_offset;
if (process_qualifier_constant(state, &loc, "xfb_offset",
qual->offset, &xfb_offset)) {
fields[i].offset = xfb_offset;
block_xfb_offset = fields[i].offset +
4 * field_type->component_slots();
}
} else {
if (layout && layout->flags.q.explicit_xfb_offset) {
unsigned align = field_type->is_64bit() ? 8 : 4;
fields[i].offset = glsl_align(block_xfb_offset, align);
block_xfb_offset += 4 * field_type->component_slots();
}
}
/* Propogate row- / column-major information down the fields of the
* structure or interface block. Structures need this data because
* the structure may contain a structure that contains ... a matrix
* that need the proper layout.
*/
if (is_interface && layout &&
(layout->flags.q.uniform || layout->flags.q.buffer) &&
(field_type->without_array()->is_matrix()
|| field_type->without_array()->is_struct())) {
/* If no layout is specified for the field, inherit the layout
* from the block.
*/
fields[i].matrix_layout = matrix_layout;
if (qual->flags.q.row_major)
fields[i].matrix_layout = GLSL_MATRIX_LAYOUT_ROW_MAJOR;
else if (qual->flags.q.column_major)
fields[i].matrix_layout = GLSL_MATRIX_LAYOUT_COLUMN_MAJOR;
/* If we're processing an uniform or buffer block, the matrix
* layout must be decided by this point.
*/
assert(fields[i].matrix_layout == GLSL_MATRIX_LAYOUT_ROW_MAJOR
|| fields[i].matrix_layout == GLSL_MATRIX_LAYOUT_COLUMN_MAJOR);
}
/* Memory qualifiers are allowed on buffer and image variables, while
* the format qualifier is only accepted for images.
*/
if (var_mode == ir_var_shader_storage ||
field_type->without_array()->is_image()) {
/* For readonly and writeonly qualifiers the field definition,
* if set, overwrites the layout qualifier.
*/
if (qual->flags.q.read_only || qual->flags.q.write_only) {
fields[i].memory_read_only = qual->flags.q.read_only;
fields[i].memory_write_only = qual->flags.q.write_only;
} else {
fields[i].memory_read_only =
layout ? layout->flags.q.read_only : 0;
fields[i].memory_write_only =
layout ? layout->flags.q.write_only : 0;
}
/* For other qualifiers, we set the flag if either the layout
* qualifier or the field qualifier are set
*/
fields[i].memory_coherent = qual->flags.q.coherent ||
(layout && layout->flags.q.coherent);
fields[i].memory_volatile = qual->flags.q._volatile ||
(layout && layout->flags.q._volatile);
fields[i].memory_restrict = qual->flags.q.restrict_flag ||
(layout && layout->flags.q.restrict_flag);
if (field_type->without_array()->is_image()) {
if (qual->flags.q.explicit_image_format) {
if (qual->image_base_type !=
field_type->without_array()->sampled_type) {
_mesa_glsl_error(&loc, state, "format qualifier doesn't "
"match the base data type of the image");
}
fields[i].image_format = qual->image_format;
} else {
if (!qual->flags.q.write_only) {
_mesa_glsl_error(&loc, state, "image not qualified with "
"`writeonly' must have a format layout "
"qualifier");
}
fields[i].image_format = PIPE_FORMAT_NONE;
}
}
}
/* Precision qualifiers do not hold any meaning in Desktop GLSL */
if (state->es_shader) {
fields[i].precision = select_gles_precision(qual->precision,
field_type,
state,
&loc);
} else {
fields[i].precision = qual->precision;
}
i++;
}
}
assert(i == decl_count);
*fields_ret = fields;
return decl_count;
}
ir_rvalue *
ast_struct_specifier::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
YYLTYPE loc = this->get_location();
unsigned expl_location = 0;
if (layout && layout->flags.q.explicit_location) {
if (!process_qualifier_constant(state, &loc, "location",
layout->location, &expl_location)) {
return NULL;
} else {
expl_location = VARYING_SLOT_VAR0 + expl_location;
}
}
glsl_struct_field *fields;
unsigned decl_count =
ast_process_struct_or_iface_block_members(instructions,
state,
&this->declarations,
&fields,
false,
GLSL_MATRIX_LAYOUT_INHERITED,
false /* allow_reserved_names */,
ir_var_auto,
layout,
0, /* for interface only */
0, /* for interface only */
0, /* for interface only */
expl_location,
0 /* for interface only */);
validate_identifier(this->name, loc, state);
type = glsl_type::get_struct_instance(fields, decl_count, this->name);
if (!type->is_anonymous() && !state->symbols->add_type(name, type)) {
const glsl_type *match = state->symbols->get_type(name);
/* allow struct matching for desktop GL - older UE4 does this */
if (match != NULL && state->is_version(130, 0) && match->record_compare(type, true, false))
_mesa_glsl_warning(& loc, state, "struct `%s' previously defined", name);
else
_mesa_glsl_error(& loc, state, "struct `%s' previously defined", name);
} else {
const glsl_type **s = reralloc(state, state->user_structures,
const glsl_type *,
state->num_user_structures + 1);
if (s != NULL) {
s[state->num_user_structures] = type;
state->user_structures = s;
state->num_user_structures++;
ir_typedecl_statement* stmt = new(state) ir_typedecl_statement(type);
/* Push the struct declarations to the top.
* However, do not insert declarations before default precision
* statements or other declarations
*/
ir_instruction* before_node = (ir_instruction*)instructions->get_head();
while (before_node &&
(before_node->ir_type == ir_type_precision ||
before_node->ir_type == ir_type_typedecl))
before_node = (ir_instruction*)before_node->next;
if (before_node)
before_node->insert_before(stmt);
else
instructions->push_head(stmt);
}
}
/* Structure type definitions do not have r-values.
*/
return NULL;
}
/**
* Visitor class which detects whether a given interface block has been used.
*/
class interface_block_usage_visitor : public ir_hierarchical_visitor
{
public:
interface_block_usage_visitor(ir_variable_mode mode, const glsl_type *block)
: mode(mode), block(block), found(false)
{
}
virtual ir_visitor_status visit(ir_dereference_variable *ir)
{
if (ir->var->data.mode == mode && ir->var->get_interface_type() == block) {
found = true;
return visit_stop;
}
return visit_continue;
}
bool usage_found() const
{
return this->found;
}
private:
ir_variable_mode mode;
const glsl_type *block;
bool found;
};
static bool
is_unsized_array_last_element(ir_variable *v)
{
const glsl_type *interface_type = v->get_interface_type();
int length = interface_type->length;
assert(v->type->is_unsized_array());
/* Check if it is the last element of the interface */
if (strcmp(interface_type->fields.structure[length-1].name, v->name) == 0)
return true;
return false;
}
static void
apply_memory_qualifiers(ir_variable *var, glsl_struct_field field)
{
var->data.memory_read_only = field.memory_read_only;
var->data.memory_write_only = field.memory_write_only;
var->data.memory_coherent = field.memory_coherent;
var->data.memory_volatile = field.memory_volatile;
var->data.memory_restrict = field.memory_restrict;
}
ir_rvalue *
ast_interface_block::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
YYLTYPE loc = this->get_location();
/* Interface blocks must be declared at global scope */
if (state->current_function != NULL) {
_mesa_glsl_error(&loc, state,
"Interface block `%s' must be declared "
"at global scope",
this->block_name);
}
/* Validate qualifiers:
*
* - Layout Qualifiers as per the table in Section 4.4
* ("Layout Qualifiers") of the GLSL 4.50 spec.
*
* - Memory Qualifiers as per Section 4.10 ("Memory Qualifiers") of the
* GLSL 4.50 spec:
*
* "Additionally, memory qualifiers may also be used in the declaration
* of shader storage blocks"
*
* Note the table in Section 4.4 says std430 is allowed on both uniform and
* buffer blocks however Section 4.4.5 (Uniform and Shader Storage Block
* Layout Qualifiers) of the GLSL 4.50 spec says:
*
* "The std430 qualifier is supported only for shader storage blocks;
* using std430 on a uniform block will result in a compile-time error."
*/
ast_type_qualifier allowed_blk_qualifiers;
allowed_blk_qualifiers.flags.i = 0;
if (this->layout.flags.q.buffer || this->layout.flags.q.uniform) {
allowed_blk_qualifiers.flags.q.shared = 1;
allowed_blk_qualifiers.flags.q.packed = 1;
allowed_blk_qualifiers.flags.q.std140 = 1;
allowed_blk_qualifiers.flags.q.row_major = 1;
allowed_blk_qualifiers.flags.q.column_major = 1;
allowed_blk_qualifiers.flags.q.explicit_align = 1;
allowed_blk_qualifiers.flags.q.explicit_binding = 1;
if (this->layout.flags.q.buffer) {
allowed_blk_qualifiers.flags.q.buffer = 1;
allowed_blk_qualifiers.flags.q.std430 = 1;
allowed_blk_qualifiers.flags.q.coherent = 1;
allowed_blk_qualifiers.flags.q._volatile = 1;
allowed_blk_qualifiers.flags.q.restrict_flag = 1;
allowed_blk_qualifiers.flags.q.read_only = 1;
allowed_blk_qualifiers.flags.q.write_only = 1;
} else {
allowed_blk_qualifiers.flags.q.uniform = 1;
}
} else {
/* Interface block */
assert(this->layout.flags.q.in || this->layout.flags.q.out);
allowed_blk_qualifiers.flags.q.explicit_location = 1;
if (this->layout.flags.q.out) {
allowed_blk_qualifiers.flags.q.out = 1;
if (state->stage == MESA_SHADER_GEOMETRY ||
state->stage == MESA_SHADER_TESS_CTRL ||
state->stage == MESA_SHADER_TESS_EVAL ||
state->stage == MESA_SHADER_VERTEX ) {
allowed_blk_qualifiers.flags.q.explicit_xfb_offset = 1;
allowed_blk_qualifiers.flags.q.explicit_xfb_buffer = 1;
allowed_blk_qualifiers.flags.q.xfb_buffer = 1;
allowed_blk_qualifiers.flags.q.explicit_xfb_stride = 1;
allowed_blk_qualifiers.flags.q.xfb_stride = 1;
if (state->stage == MESA_SHADER_GEOMETRY) {
allowed_blk_qualifiers.flags.q.stream = 1;
allowed_blk_qualifiers.flags.q.explicit_stream = 1;
}
if (state->stage == MESA_SHADER_TESS_CTRL) {
allowed_blk_qualifiers.flags.q.patch = 1;
}
}
} else {
allowed_blk_qualifiers.flags.q.in = 1;
if (state->stage == MESA_SHADER_TESS_EVAL) {
allowed_blk_qualifiers.flags.q.patch = 1;
}
}
}
this->layout.validate_flags(&loc, state, allowed_blk_qualifiers,
"invalid qualifier for block",
this->block_name);
enum glsl_interface_packing packing;
if (this->layout.flags.q.std140) {
packing = GLSL_INTERFACE_PACKING_STD140;
} else if (this->layout.flags.q.packed) {
packing = GLSL_INTERFACE_PACKING_PACKED;
} else if (this->layout.flags.q.std430) {
packing = GLSL_INTERFACE_PACKING_STD430;
} else {
/* The default layout is shared.
*/
packing = GLSL_INTERFACE_PACKING_SHARED;
}
ir_variable_mode var_mode;
const char *iface_type_name;
if (this->layout.flags.q.in) {
var_mode = ir_var_shader_in;
iface_type_name = "in";
} else if (this->layout.flags.q.out) {
var_mode = ir_var_shader_out;
iface_type_name = "out";
} else if (this->layout.flags.q.uniform) {
var_mode = ir_var_uniform;
iface_type_name = "uniform";
} else if (this->layout.flags.q.buffer) {
var_mode = ir_var_shader_storage;
iface_type_name = "buffer";
} else {
var_mode = ir_var_auto;
iface_type_name = "UNKNOWN";
assert(!"interface block layout qualifier not found!");
}
enum glsl_matrix_layout matrix_layout = GLSL_MATRIX_LAYOUT_INHERITED;
if (this->layout.flags.q.row_major)
matrix_layout = GLSL_MATRIX_LAYOUT_ROW_MAJOR;
else if (this->layout.flags.q.column_major)
matrix_layout = GLSL_MATRIX_LAYOUT_COLUMN_MAJOR;
bool redeclaring_per_vertex = strcmp(this->block_name, "gl_PerVertex") == 0;
exec_list declared_variables;
glsl_struct_field *fields;
/* For blocks that accept memory qualifiers (i.e. shader storage), verify
* that we don't have incompatible qualifiers
*/
if (this->layout.flags.q.read_only && this->layout.flags.q.write_only) {
_mesa_glsl_error(&loc, state,
"Interface block sets both readonly and writeonly");
}
unsigned qual_stream;
if (!process_qualifier_constant(state, &loc, "stream", this->layout.stream,
&qual_stream) ||
!validate_stream_qualifier(&loc, state, qual_stream)) {
/* If the stream qualifier is invalid it doesn't make sense to continue
* on and try to compare stream layouts on member variables against it
* so just return early.
*/
return NULL;
}
unsigned qual_xfb_buffer;
if (!process_qualifier_constant(state, &loc, "xfb_buffer",
layout.xfb_buffer, &qual_xfb_buffer) ||
!validate_xfb_buffer_qualifier(&loc, state, qual_xfb_buffer)) {
return NULL;
}
unsigned qual_xfb_offset;
if (layout.flags.q.explicit_xfb_offset) {
if (!process_qualifier_constant(state, &loc, "xfb_offset",
layout.offset, &qual_xfb_offset)) {
return NULL;
}
}
unsigned qual_xfb_stride;
if (layout.flags.q.explicit_xfb_stride) {
if (!process_qualifier_constant(state, &loc, "xfb_stride",
layout.xfb_stride, &qual_xfb_stride)) {
return NULL;
}
}
unsigned expl_location = 0;
if (layout.flags.q.explicit_location) {
if (!process_qualifier_constant(state, &loc, "location",
layout.location, &expl_location)) {
return NULL;
} else {
expl_location += this->layout.flags.q.patch ? VARYING_SLOT_PATCH0
: VARYING_SLOT_VAR0;
}
}
unsigned expl_align = 0;
if (layout.flags.q.explicit_align) {
if (!process_qualifier_constant(state, &loc, "align",
layout.align, &expl_align)) {
return NULL;
} else {
if (expl_align == 0 || expl_align & (expl_align - 1)) {
_mesa_glsl_error(&loc, state, "align layout qualifier is not a "
"power of 2.");
return NULL;
}
}
}
unsigned int num_variables =
ast_process_struct_or_iface_block_members(&declared_variables,
state,
&this->declarations,
&fields,
true,
matrix_layout,
redeclaring_per_vertex,
var_mode,
&this->layout,
qual_stream,
qual_xfb_buffer,
qual_xfb_offset,
expl_location,
expl_align);
if (!redeclaring_per_vertex) {
validate_identifier(this->block_name, loc, state);
/* From section 4.3.9 ("Interface Blocks") of the GLSL 4.50 spec:
*
* "Block names have no other use within a shader beyond interface
* matching; it is a compile-time error to use a block name at global
* scope for anything other than as a block name."
*/
ir_variable *var = state->symbols->get_variable(this->block_name);
if (var && !var->type->is_interface()) {
_mesa_glsl_error(&loc, state, "Block name `%s' is "
"already used in the scope.",
this->block_name);
}
}
const glsl_type *earlier_per_vertex = NULL;
if (redeclaring_per_vertex) {
/* Find the previous declaration of gl_PerVertex. If we're redeclaring
* the named interface block gl_in, we can find it by looking at the
* previous declaration of gl_in. Otherwise we can find it by looking
* at the previous decalartion of any of the built-in outputs,
* e.g. gl_Position.
*
* Also check that the instance name and array-ness of the redeclaration
* are correct.
*/
switch (var_mode) {
case ir_var_shader_in:
if (ir_variable *earlier_gl_in =
state->symbols->get_variable("gl_in")) {
earlier_per_vertex = earlier_gl_in->get_interface_type();
} else {
_mesa_glsl_error(&loc, state,
"redeclaration of gl_PerVertex input not allowed "
"in the %s shader",
_mesa_shader_stage_to_string(state->stage));
}
if (this->instance_name == NULL ||
strcmp(this->instance_name, "gl_in") != 0 || this->array_specifier == NULL ||
!this->array_specifier->is_single_dimension()) {
_mesa_glsl_error(&loc, state,
"gl_PerVertex input must be redeclared as "
"gl_in[]");
}
break;
case ir_var_shader_out:
if (ir_variable *earlier_gl_Position =
state->symbols->get_variable("gl_Position")) {
earlier_per_vertex = earlier_gl_Position->get_interface_type();
} else if (ir_variable *earlier_gl_out =
state->symbols->get_variable("gl_out")) {
earlier_per_vertex = earlier_gl_out->get_interface_type();
} else {
_mesa_glsl_error(&loc, state,
"redeclaration of gl_PerVertex output not "
"allowed in the %s shader",
_mesa_shader_stage_to_string(state->stage));
}
if (state->stage == MESA_SHADER_TESS_CTRL) {
if (this->instance_name == NULL ||
strcmp(this->instance_name, "gl_out") != 0 || this->array_specifier == NULL) {
_mesa_glsl_error(&loc, state,
"gl_PerVertex output must be redeclared as "
"gl_out[]");
}
} else {
if (this->instance_name != NULL) {
_mesa_glsl_error(&loc, state,
"gl_PerVertex output may not be redeclared with "
"an instance name");
}
}
break;
default:
_mesa_glsl_error(&loc, state,
"gl_PerVertex must be declared as an input or an "
"output");
break;
}
if (earlier_per_vertex == NULL) {
/* An error has already been reported. Bail out to avoid null
* dereferences later in this function.
*/
return NULL;
}
/* Copy locations from the old gl_PerVertex interface block. */
for (unsigned i = 0; i < num_variables; i++) {
int j = earlier_per_vertex->field_index(fields[i].name);
if (j == -1) {
_mesa_glsl_error(&loc, state,
"redeclaration of gl_PerVertex must be a subset "
"of the built-in members of gl_PerVertex");
} else {
fields[i].location =
earlier_per_vertex->fields.structure[j].location;
fields[i].offset =
earlier_per_vertex->fields.structure[j].offset;
fields[i].interpolation =
earlier_per_vertex->fields.structure[j].interpolation;
fields[i].centroid =
earlier_per_vertex->fields.structure[j].centroid;
fields[i].sample =
earlier_per_vertex->fields.structure[j].sample;
fields[i].patch =
earlier_per_vertex->fields.structure[j].patch;
fields[i].precision =
earlier_per_vertex->fields.structure[j].precision;
fields[i].explicit_xfb_buffer =
earlier_per_vertex->fields.structure[j].explicit_xfb_buffer;
fields[i].xfb_buffer =
earlier_per_vertex->fields.structure[j].xfb_buffer;
fields[i].xfb_stride =
earlier_per_vertex->fields.structure[j].xfb_stride;
}
}
/* From section 7.1 ("Built-in Language Variables") of the GLSL 4.10
* spec:
*
* If a built-in interface block is redeclared, it must appear in
* the shader before any use of any member included in the built-in
* declaration, or a compilation error will result.
*
* This appears to be a clarification to the behaviour established for
* gl_PerVertex by GLSL 1.50, therefore we implement this behaviour
* regardless of GLSL version.
*/
interface_block_usage_visitor v(var_mode, earlier_per_vertex);
v.run(instructions);
if (v.usage_found()) {
_mesa_glsl_error(&loc, state,
"redeclaration of a built-in interface block must "
"appear before any use of any member of the "
"interface block");
}
}
const glsl_type *block_type =
glsl_type::get_interface_instance(fields,
num_variables,
packing,
matrix_layout ==
GLSL_MATRIX_LAYOUT_ROW_MAJOR,
this->block_name);
unsigned component_size = block_type->contains_double() ? 8 : 4;
int xfb_offset =
layout.flags.q.explicit_xfb_offset ? (int) qual_xfb_offset : -1;
validate_xfb_offset_qualifier(&loc, state, xfb_offset, block_type,
component_size);
if (!state->symbols->add_interface(block_type->name, block_type, var_mode)) {
YYLTYPE loc = this->get_location();
_mesa_glsl_error(&loc, state, "interface block `%s' with type `%s' "
"already taken in the current scope",
this->block_name, iface_type_name);
}
/* Since interface blocks cannot contain statements, it should be
* impossible for the block to generate any instructions.
*/
assert(declared_variables.is_empty());
/* From section 4.3.4 (Inputs) of the GLSL 1.50 spec:
*
* Geometry shader input variables get the per-vertex values written
* out by vertex shader output variables of the same names. Since a
* geometry shader operates on a set of vertices, each input varying
* variable (or input block, see interface blocks below) needs to be
* declared as an array.
*/
if (state->stage == MESA_SHADER_GEOMETRY && this->array_specifier == NULL &&
var_mode == ir_var_shader_in) {
_mesa_glsl_error(&loc, state, "geometry shader inputs must be arrays");
} else if ((state->stage == MESA_SHADER_TESS_CTRL ||
state->stage == MESA_SHADER_TESS_EVAL) &&
!this->layout.flags.q.patch &&
this->array_specifier == NULL &&
var_mode == ir_var_shader_in) {
_mesa_glsl_error(&loc, state, "per-vertex tessellation shader inputs must be arrays");
} else if (state->stage == MESA_SHADER_TESS_CTRL &&
!this->layout.flags.q.patch &&
this->array_specifier == NULL &&
var_mode == ir_var_shader_out) {
_mesa_glsl_error(&loc, state, "tessellation control shader outputs must be arrays");
}
ir_typedecl_statement* stmt = new(state) ir_typedecl_statement(block_type);
/* Push the interface declarations to the top.
* However, do not insert declarations before default precision
* statements or other declarations
*/
ir_instruction* before_node = (ir_instruction*)instructions->get_head();
while (before_node &&
(before_node->ir_type == ir_type_precision ||
before_node->ir_type == ir_type_typedecl))
before_node = (ir_instruction*)before_node->next;
if (before_node)
before_node->insert_before(stmt);
else
instructions->push_head(stmt);
/* Page 39 (page 45 of the PDF) of section 4.3.7 in the GLSL ES 3.00 spec
* says:
*
* "If an instance name (instance-name) is used, then it puts all the
* members inside a scope within its own name space, accessed with the
* field selector ( . ) operator (analogously to structures)."
*/
if (this->instance_name) {
if (redeclaring_per_vertex) {
/* When a built-in in an unnamed interface block is redeclared,
* get_variable_being_redeclared() calls
* check_builtin_array_max_size() to make sure that built-in array
* variables aren't redeclared to illegal sizes. But we're looking
* at a redeclaration of a named built-in interface block. So we
* have to manually call check_builtin_array_max_size() for all parts
* of the interface that are arrays.
*/
for (unsigned i = 0; i < num_variables; i++) {
if (fields[i].type->is_array()) {
const unsigned size = fields[i].type->array_size();
check_builtin_array_max_size(fields[i].name, size, loc, state);
}
}
} else {
validate_identifier(this->instance_name, loc, state);
}
ir_variable *var;
if (this->array_specifier != NULL) {
const glsl_type *block_array_type =
process_array_type(&loc, block_type, this->array_specifier, state);
/* Section 4.3.7 (Interface Blocks) of the GLSL 1.50 spec says:
*
* For uniform blocks declared an array, each individual array
* element corresponds to a separate buffer object backing one
* instance of the block. As the array size indicates the number
* of buffer objects needed, uniform block array declarations
* must specify an array size.
*
* And a few paragraphs later:
*
* Geometry shader input blocks must be declared as arrays and
* follow the array declaration and linking rules for all
* geometry shader inputs. All other input and output block
* arrays must specify an array size.
*
* The same applies to tessellation shaders.
*
* The upshot of this is that the only circumstance where an
* interface array size *doesn't* need to be specified is on a
* geometry shader input, tessellation control shader input,
* tessellation control shader output, and tessellation evaluation
* shader input.
*/
if (block_array_type->is_unsized_array()) {
bool allow_inputs = state->stage == MESA_SHADER_GEOMETRY ||
state->stage == MESA_SHADER_TESS_CTRL ||
state->stage == MESA_SHADER_TESS_EVAL;
bool allow_outputs = state->stage == MESA_SHADER_TESS_CTRL;
if (this->layout.flags.q.in) {
if (!allow_inputs)
_mesa_glsl_error(&loc, state,
"unsized input block arrays not allowed in "
"%s shader",
_mesa_shader_stage_to_string(state->stage));
} else if (this->layout.flags.q.out) {
if (!allow_outputs)
_mesa_glsl_error(&loc, state,
"unsized output block arrays not allowed in "
"%s shader",
_mesa_shader_stage_to_string(state->stage));
} else {
/* by elimination, this is a uniform block array */
_mesa_glsl_error(&loc, state,
"unsized uniform block arrays not allowed in "
"%s shader",
_mesa_shader_stage_to_string(state->stage));
}
}
/* From section 4.3.9 (Interface Blocks) of the GLSL ES 3.10 spec:
*
* * Arrays of arrays of blocks are not allowed
*/
if (state->es_shader && block_array_type->is_array() &&
block_array_type->fields.array->is_array()) {
_mesa_glsl_error(&loc, state,
"arrays of arrays interface blocks are "
"not allowed");
}
var = new(state) ir_variable(block_array_type,
this->instance_name,
var_mode);
} else {
var = new(state) ir_variable(block_type,
this->instance_name,
var_mode);
}
var->data.matrix_layout = matrix_layout == GLSL_MATRIX_LAYOUT_INHERITED
? GLSL_MATRIX_LAYOUT_COLUMN_MAJOR : matrix_layout;
if (var_mode == ir_var_shader_in || var_mode == ir_var_uniform)
var->data.read_only = true;
var->data.patch = this->layout.flags.q.patch;
if (state->stage == MESA_SHADER_GEOMETRY && var_mode == ir_var_shader_in)
handle_geometry_shader_input_decl(state, loc, var);
else if ((state->stage == MESA_SHADER_TESS_CTRL ||
state->stage == MESA_SHADER_TESS_EVAL) && var_mode == ir_var_shader_in)
handle_tess_shader_input_decl(state, loc, var);
else if (state->stage == MESA_SHADER_TESS_CTRL && var_mode == ir_var_shader_out)
handle_tess_ctrl_shader_output_decl(state, loc, var);
for (unsigned i = 0; i < num_variables; i++) {
if (var->data.mode == ir_var_shader_storage)
apply_memory_qualifiers(var, fields[i]);
}
if (ir_variable *earlier =
state->symbols->get_variable(this->instance_name)) {
if (!redeclaring_per_vertex) {
_mesa_glsl_error(&loc, state, "`%s' redeclared",
this->instance_name);
}
earlier->data.how_declared = ir_var_declared_normally;
earlier->type = var->type;
earlier->reinit_interface_type(block_type);
delete var;
} else {
if (this->layout.flags.q.explicit_binding) {
apply_explicit_binding(state, &loc, var, var->type,
&this->layout);
}
var->data.stream = qual_stream;
if (layout.flags.q.explicit_location) {
var->data.location = expl_location;
var->data.explicit_location = true;
}
state->symbols->add_variable(var);
instructions->push_tail(var);
}
} else {
/* In order to have an array size, the block must also be declared with
* an instance name.
*/
assert(this->array_specifier == NULL);
for (unsigned i = 0; i < num_variables; i++) {
ir_variable *var =
new(state) ir_variable(fields[i].type,
ralloc_strdup(state, fields[i].name),
var_mode);
var->data.interpolation = fields[i].interpolation;
var->data.centroid = fields[i].centroid;
var->data.sample = fields[i].sample;
var->data.patch = fields[i].patch;
var->data.stream = qual_stream;
var->data.location = fields[i].location;
if (fields[i].location != -1)
var->data.explicit_location = true;
var->data.explicit_xfb_buffer = fields[i].explicit_xfb_buffer;
var->data.xfb_buffer = fields[i].xfb_buffer;
if (fields[i].offset != -1)
var->data.explicit_xfb_offset = true;
var->data.offset = fields[i].offset;
var->init_interface_type(block_type);
if (var_mode == ir_var_shader_in || var_mode == ir_var_uniform)
var->data.read_only = true;
/* Precision qualifiers do not have any meaning in Desktop GLSL */
if (state->es_shader) {
var->data.precision =
select_gles_precision(fields[i].precision, fields[i].type,
state, &loc);
}
if (fields[i].matrix_layout == GLSL_MATRIX_LAYOUT_INHERITED) {
var->data.matrix_layout = matrix_layout == GLSL_MATRIX_LAYOUT_INHERITED
? GLSL_MATRIX_LAYOUT_COLUMN_MAJOR : matrix_layout;
} else {
var->data.matrix_layout = fields[i].matrix_layout;
}
if (var->data.mode == ir_var_shader_storage)
apply_memory_qualifiers(var, fields[i]);
/* Examine var name here since var may get deleted in the next call */
bool var_is_gl_id = is_gl_identifier(var->name);
if (redeclaring_per_vertex) {
bool is_redeclaration;
var =
get_variable_being_redeclared(&var, loc, state,
true /* allow_all_redeclarations */,
&is_redeclaration);
if (!var_is_gl_id || !is_redeclaration) {
_mesa_glsl_error(&loc, state,
"redeclaration of gl_PerVertex can only "
"include built-in variables");
} else if (var->data.how_declared == ir_var_declared_normally) {
_mesa_glsl_error(&loc, state,
"`%s' has already been redeclared",
var->name);
} else {
var->data.how_declared = ir_var_declared_in_block;
var->reinit_interface_type(block_type);
}
continue;
}
if (state->symbols->get_variable(var->name) != NULL)
_mesa_glsl_error(&loc, state, "`%s' redeclared", var->name);
/* Propagate the "binding" keyword into this UBO/SSBO's fields.
* The UBO declaration itself doesn't get an ir_variable unless it
* has an instance name. This is ugly.
*/
if (this->layout.flags.q.explicit_binding) {
apply_explicit_binding(state, &loc, var,
var->get_interface_type(), &this->layout);
}
if (var->type->is_unsized_array()) {
if (var->is_in_shader_storage_block() &&
is_unsized_array_last_element(var)) {
var->data.from_ssbo_unsized_array = true;
} else {
/* From GLSL ES 3.10 spec, section 4.1.9 "Arrays":
*
* "If an array is declared as the last member of a shader storage
* block and the size is not specified at compile-time, it is
* sized at run-time. In all other cases, arrays are sized only
* at compile-time."
*
* In desktop GLSL it is allowed to have unsized-arrays that are
* not last, as long as we can determine that they are implicitly
* sized.
*/
if (state->es_shader) {
_mesa_glsl_error(&loc, state, "unsized array `%s' "
"definition: only last member of a shader "
"storage block can be defined as unsized "
"array", fields[i].name);
}
}
}
state->symbols->add_variable(var);
instructions->push_tail(var);
}
if (redeclaring_per_vertex && block_type != earlier_per_vertex) {
/* From section 7.1 ("Built-in Language Variables") of the GLSL 4.10 spec:
*
* It is also a compilation error ... to redeclare a built-in
* block and then use a member from that built-in block that was
* not included in the redeclaration.
*
* This appears to be a clarification to the behaviour established
* for gl_PerVertex by GLSL 1.50, therefore we implement this
* behaviour regardless of GLSL version.
*
* To prevent the shader from using a member that was not included in
* the redeclaration, we disable any ir_variables that are still
* associated with the old declaration of gl_PerVertex (since we've
* already updated all of the variables contained in the new
* gl_PerVertex to point to it).
*
* As a side effect this will prevent
* validate_intrastage_interface_blocks() from getting confused and
* thinking there are conflicting definitions of gl_PerVertex in the
* shader.
*/
foreach_in_list_safe(ir_instruction, node, instructions) {
ir_variable *const var = node->as_variable();
if (var != NULL &&
var->get_interface_type() == earlier_per_vertex &&
var->data.mode == var_mode) {
if (var->data.how_declared == ir_var_declared_normally) {
_mesa_glsl_error(&loc, state,
"redeclaration of gl_PerVertex cannot "
"follow a redeclaration of `%s'",
var->name);
}
state->symbols->disable_variable(var->name);
var->remove();
}
}
}
}
return NULL;
}
ir_rvalue *
ast_tcs_output_layout::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
YYLTYPE loc = this->get_location();
unsigned num_vertices;
if (!state->out_qualifier->vertices->
process_qualifier_constant(state, "vertices", &num_vertices,
false)) {
/* return here to stop cascading incorrect error messages */
return NULL;
}
/* If any shader outputs occurred before this declaration and specified an
* array size, make sure the size they specified is consistent with the
* primitive type.
*/
if (state->tcs_output_size != 0 && state->tcs_output_size != num_vertices) {
_mesa_glsl_error(&loc, state,
"this tessellation control shader output layout "
"specifies %u vertices, but a previous output "
"is declared with size %u",
num_vertices, state->tcs_output_size);
return NULL;
}
state->tcs_output_vertices_specified = true;
/* If any shader outputs occurred before this declaration and did not
* specify an array size, their size is determined now.
*/
foreach_in_list (ir_instruction, node, instructions) {
ir_variable *var = node->as_variable();
if (var == NULL || var->data.mode != ir_var_shader_out)
continue;
/* Note: Not all tessellation control shader output are arrays. */
if (!var->type->is_unsized_array() || var->data.patch)
continue;
if (var->data.max_array_access >= (int)num_vertices) {
_mesa_glsl_error(&loc, state,
"this tessellation control shader output layout "
"specifies %u vertices, but an access to element "
"%u of output `%s' already exists", num_vertices,
var->data.max_array_access, var->name);
} else {
var->type = glsl_type::get_array_instance(var->type->fields.array,
num_vertices);
}
}
return NULL;
}
ir_rvalue *
ast_gs_input_layout::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
YYLTYPE loc = this->get_location();
/* Should have been prevented by the parser. */
assert(!state->gs_input_prim_type_specified
|| state->in_qualifier->prim_type == this->prim_type);
/* If any shader inputs occurred before this declaration and specified an
* array size, make sure the size they specified is consistent with the
* primitive type.
*/
unsigned num_vertices = vertices_per_prim(this->prim_type);
if (state->gs_input_size != 0 && state->gs_input_size != num_vertices) {
_mesa_glsl_error(&loc, state,
"this geometry shader input layout implies %u vertices"
" per primitive, but a previous input is declared"
" with size %u", num_vertices, state->gs_input_size);
return NULL;
}
state->gs_input_prim_type_specified = true;
/* If any shader inputs occurred before this declaration and did not
* specify an array size, their size is determined now.
*/
foreach_in_list(ir_instruction, node, instructions) {
ir_variable *var = node->as_variable();
if (var == NULL || var->data.mode != ir_var_shader_in)
continue;
/* Note: gl_PrimitiveIDIn has mode ir_var_shader_in, but it's not an
* array; skip it.
*/
if (var->type->is_unsized_array()) {
if (var->data.max_array_access >= (int)num_vertices) {
_mesa_glsl_error(&loc, state,
"this geometry shader input layout implies %u"
" vertices, but an access to element %u of input"
" `%s' already exists", num_vertices,
var->data.max_array_access, var->name);
} else {
var->type = glsl_type::get_array_instance(var->type->fields.array,
num_vertices);
}
}
}
return NULL;
}
ir_rvalue *
ast_cs_input_layout::hir(exec_list *instructions,
struct _mesa_glsl_parse_state *state)
{
YYLTYPE loc = this->get_location();
/* From the ARB_compute_shader specification:
*
* If the local size of the shader in any dimension is greater
* than the maximum size supported by the implementation for that
* dimension, a compile-time error results.
*
* It is not clear from the spec how the error should be reported if
* the total size of the work group exceeds
* MAX_COMPUTE_WORK_GROUP_INVOCATIONS, but it seems reasonable to
* report it at compile time as well.
*/
GLuint64 total_invocations = 1;
unsigned qual_local_size[3];
for (int i = 0; i < 3; i++) {
char *local_size_str = ralloc_asprintf(NULL, "invalid local_size_%c",
'x' + i);
/* Infer a local_size of 1 for unspecified dimensions */
if (this->local_size[i] == NULL) {
qual_local_size[i] = 1;
} else if (!this->local_size[i]->
process_qualifier_constant(state, local_size_str,
&qual_local_size[i], false)) {
ralloc_free(local_size_str);
return NULL;
}
ralloc_free(local_size_str);
if (qual_local_size[i] > state->ctx->Const.MaxComputeWorkGroupSize[i]) {
_mesa_glsl_error(&loc, state,
"local_size_%c exceeds MAX_COMPUTE_WORK_GROUP_SIZE"
" (%d)", 'x' + i,
state->ctx->Const.MaxComputeWorkGroupSize[i]);
break;
}
total_invocations *= qual_local_size[i];
if (total_invocations >
state->ctx->Const.MaxComputeWorkGroupInvocations) {
_mesa_glsl_error(&loc, state,
"product of local_sizes exceeds "
"MAX_COMPUTE_WORK_GROUP_INVOCATIONS (%d)",
state->ctx->Const.MaxComputeWorkGroupInvocations);
break;
}
}
/* If any compute input layout declaration preceded this one, make sure it
* was consistent with this one.
*/
if (state->cs_input_local_size_specified) {
for (int i = 0; i < 3; i++) {
if (state->cs_input_local_size[i] != qual_local_size[i]) {
_mesa_glsl_error(&loc, state,
"compute shader input layout does not match"
" previous declaration");
return NULL;
}
}
}
/* The ARB_compute_variable_group_size spec says:
*
* If a compute shader including a *local_size_variable* qualifier also
* declares a fixed local group size using the *local_size_x*,
* *local_size_y*, or *local_size_z* qualifiers, a compile-time error
* results
*/
if (state->cs_input_local_size_variable_specified) {
_mesa_glsl_error(&loc, state,
"compute shader can't include both a variable and a "
"fixed local group size");
return NULL;
}
state->cs_input_local_size_specified = true;
for (int i = 0; i < 3; i++)
state->cs_input_local_size[i] = qual_local_size[i];
/* We may now declare the built-in constant gl_WorkGroupSize (see
* builtin_variable_generator::generate_constants() for why we didn't
* declare it earlier).
*/
ir_variable *var = new(state->symbols)
ir_variable(glsl_type::uvec3_type, "gl_WorkGroupSize", ir_var_auto);
var->data.how_declared = ir_var_declared_implicitly;
var->data.read_only = true;
instructions->push_tail(var);
state->symbols->add_variable(var);
ir_constant_data data;
memset(&data, 0, sizeof(data));
for (int i = 0; i < 3; i++)
data.u[i] = qual_local_size[i];
var->constant_value = new(var) ir_constant(glsl_type::uvec3_type, &data);
var->constant_initializer =
new(var) ir_constant(glsl_type::uvec3_type, &data);
var->data.has_initializer = true;
return NULL;
}
static void
detect_conflicting_assignments(struct _mesa_glsl_parse_state *state,
exec_list *instructions)
{
bool gl_FragColor_assigned = false;
bool gl_FragData_assigned = false;
bool gl_FragSecondaryColor_assigned = false;
bool gl_FragSecondaryData_assigned = false;
bool user_defined_fs_output_assigned = false;
ir_variable *user_defined_fs_output = NULL;
/* It would be nice to have proper location information. */
YYLTYPE loc;
memset(&loc, 0, sizeof(loc));
foreach_in_list(ir_instruction, node, instructions) {
ir_variable *var = node->as_variable();
if (!var || !var->data.assigned)
continue;
if (strcmp(var->name, "gl_FragColor") == 0)
gl_FragColor_assigned = true;
else if (strcmp(var->name, "gl_FragData") == 0)
gl_FragData_assigned = true;
else if (strcmp(var->name, "gl_SecondaryFragColorEXT") == 0)
gl_FragSecondaryColor_assigned = true;
else if (strcmp(var->name, "gl_SecondaryFragDataEXT") == 0)
gl_FragSecondaryData_assigned = true;
else if (!is_gl_identifier(var->name)) {
if (state->stage == MESA_SHADER_FRAGMENT &&
var->data.mode == ir_var_shader_out) {
user_defined_fs_output_assigned = true;
user_defined_fs_output = var;
}
}
}
/* From the GLSL 1.30 spec:
*
* "If a shader statically assigns a value to gl_FragColor, it
* may not assign a value to any element of gl_FragData. If a
* shader statically writes a value to any element of
* gl_FragData, it may not assign a value to
* gl_FragColor. That is, a shader may assign values to either
* gl_FragColor or gl_FragData, but not both. Multiple shaders
* linked together must also consistently write just one of
* these variables. Similarly, if user declared output
* variables are in use (statically assigned to), then the
* built-in variables gl_FragColor and gl_FragData may not be
* assigned to. These incorrect usages all generate compile
* time errors."
*/
if (gl_FragColor_assigned && gl_FragData_assigned) {
_mesa_glsl_error(&loc, state, "fragment shader writes to both "
"`gl_FragColor' and `gl_FragData'");
} else if (gl_FragColor_assigned && user_defined_fs_output_assigned) {
_mesa_glsl_error(&loc, state, "fragment shader writes to both "
"`gl_FragColor' and `%s'",
user_defined_fs_output->name);
} else if (gl_FragSecondaryColor_assigned && gl_FragSecondaryData_assigned) {
_mesa_glsl_error(&loc, state, "fragment shader writes to both "
"`gl_FragSecondaryColorEXT' and"
" `gl_FragSecondaryDataEXT'");
} else if (gl_FragColor_assigned && gl_FragSecondaryData_assigned) {
_mesa_glsl_error(&loc, state, "fragment shader writes to both "
"`gl_FragColor' and"
" `gl_FragSecondaryDataEXT'");
} else if (gl_FragData_assigned && gl_FragSecondaryColor_assigned) {
_mesa_glsl_error(&loc, state, "fragment shader writes to both "
"`gl_FragData' and"
" `gl_FragSecondaryColorEXT'");
} else if (gl_FragData_assigned && user_defined_fs_output_assigned) {
_mesa_glsl_error(&loc, state, "fragment shader writes to both "
"`gl_FragData' and `%s'",
user_defined_fs_output->name);
}
if ((gl_FragSecondaryColor_assigned || gl_FragSecondaryData_assigned) &&
!state->EXT_blend_func_extended_enable) {
_mesa_glsl_error(&loc, state,
"Dual source blending requires EXT_blend_func_extended");
}
}
static void
verify_subroutine_associated_funcs(struct _mesa_glsl_parse_state *state)
{
YYLTYPE loc;
memset(&loc, 0, sizeof(loc));
/* Section 6.1.2 (Subroutines) of the GLSL 4.00 spec says:
*
* "A program will fail to compile or link if any shader
* or stage contains two or more functions with the same
* name if the name is associated with a subroutine type."
*/
for (int i = 0; i < state->num_subroutines; i++) {
unsigned definitions = 0;
ir_function *fn = state->subroutines[i];
/* Calculate number of function definitions with the same name */
foreach_in_list(ir_function_signature, sig, &fn->signatures) {
if (sig->is_defined) {
if (++definitions > 1) {
_mesa_glsl_error(&loc, state,
"%s shader contains two or more function "
"definitions with name `%s', which is "
"associated with a subroutine type.\n",
_mesa_shader_stage_to_string(state->stage),
fn->name);
return;
}
}
}
}
}
static void
remove_per_vertex_blocks(exec_list *instructions,
_mesa_glsl_parse_state *state, ir_variable_mode mode)
{
/* Find the gl_PerVertex interface block of the appropriate (in/out) mode,
* if it exists in this shader type.
*/
const glsl_type *per_vertex = NULL;
switch (mode) {
case ir_var_shader_in:
if (ir_variable *gl_in = state->symbols->get_variable("gl_in"))
per_vertex = gl_in->get_interface_type();
break;
case ir_var_shader_out:
if (ir_variable *gl_Position =
state->symbols->get_variable("gl_Position")) {
per_vertex = gl_Position->get_interface_type();
}
break;
default:
assert(!"Unexpected mode");
break;
}
/* If we didn't find a built-in gl_PerVertex interface block, then we don't
* need to do anything.
*/
if (per_vertex == NULL)
return;
/* If the interface block is used by the shader, then we don't need to do
* anything.
*/
interface_block_usage_visitor v(mode, per_vertex);
v.run(instructions);
if (v.usage_found())
return;
/* Remove any ir_variable declarations that refer to the interface block
* we're removing.
*/
foreach_in_list_safe(ir_instruction, node, instructions) {
ir_variable *const var = node->as_variable();
if (var != NULL && var->get_interface_type() == per_vertex &&
var->data.mode == mode) {
state->symbols->disable_variable(var->name);
var->remove();
}
}
}
ir_rvalue *
ast_warnings_toggle::hir(exec_list *,
struct _mesa_glsl_parse_state *state)
{
state->warnings_enabled = enable;
return NULL;
}