Remove redundant test_
for test functions
This commit is contained in:
parent
85ea81a41d
commit
b8799233f8
10
src/lib.rs
10
src/lib.rs
|
@ -172,7 +172,7 @@ mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_whitespace() {
|
fn whitespace() {
|
||||||
let source_code = " \t\n";
|
let source_code = " \t\n";
|
||||||
let mut lexer = Lexer::new(source_code);
|
let mut lexer = Lexer::new(source_code);
|
||||||
let tokens = lexer.get_tokens();
|
let tokens = lexer.get_tokens();
|
||||||
|
@ -180,7 +180,7 @@ mod tests {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_identifier() {
|
fn identifier() {
|
||||||
let source_code = "variableName";
|
let source_code = "variableName";
|
||||||
let mut lexer = Lexer::new(source_code);
|
let mut lexer = Lexer::new(source_code);
|
||||||
let tokens = lexer.get_tokens();
|
let tokens = lexer.get_tokens();
|
||||||
|
@ -194,7 +194,7 @@ mod tests {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_keyword() {
|
fn keyword() {
|
||||||
let source_code = "uniform";
|
let source_code = "uniform";
|
||||||
let mut lexer = Lexer::new(source_code);
|
let mut lexer = Lexer::new(source_code);
|
||||||
let tokens = lexer.get_tokens();
|
let tokens = lexer.get_tokens();
|
||||||
|
@ -205,7 +205,7 @@ mod tests {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_integer_literal() {
|
fn integer_literal() {
|
||||||
let source_code = "12345";
|
let source_code = "12345";
|
||||||
let mut lexer = Lexer::new(source_code);
|
let mut lexer = Lexer::new(source_code);
|
||||||
let tokens = lexer.get_tokens();
|
let tokens = lexer.get_tokens();
|
||||||
|
@ -213,7 +213,7 @@ mod tests {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_float_literal() {
|
fn float_literal() {
|
||||||
let source_code = "123.45";
|
let source_code = "123.45";
|
||||||
let mut lexer = Lexer::new(source_code);
|
let mut lexer = Lexer::new(source_code);
|
||||||
let tokens = lexer.get_tokens();
|
let tokens = lexer.get_tokens();
|
||||||
|
|
Loading…
Reference in a new issue