Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
J
json
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Analytics
Analytics
CI / CD
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
Libraries
json
Commits
c2d55109
Unverified
Commit
c2d55109
authored
Feb 11, 2017
by
Niels Lohmann
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
🔨
make lexer distinguishes number types
parent
9f5dccba
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
275 additions
and
215 deletions
+275
-215
src/json.hpp
src/json.hpp
+190
-128
src/json.hpp.re2c
src/json.hpp.re2c
+60
-76
test/src/unit-class_lexer.cpp
test/src/unit-class_lexer.cpp
+25
-11
No files found.
src/json.hpp
View file @
c2d55109
...
@@ -9444,7 +9444,9 @@ class basic_json
...
@@ -9444,7 +9444,9 @@ class basic_json
literal_false
,
///< the `false` literal
literal_false
,
///< the `false` literal
literal_null
,
///< the `null` literal
literal_null
,
///< the `null` literal
value_string
,
///< a string -- use get_string() for actual value
value_string
,
///< a string -- use get_string() for actual value
value_number
,
///< a number -- use get_number() for actual value
value_unsigned_integer
,
///< an unsigned integer -- use get_number() for actual value
value_signed_integer
,
///< a signed integer -- use get_number() for actual value
value_float
,
///< an floating point number -- use get_number() for actual value
begin_array
,
///< the character for array begin `[`
begin_array
,
///< the character for array begin `[`
begin_object
,
///< the character for object begin `{`
begin_object
,
///< the character for object begin `{`
end_array
,
///< the character for array end `]`
end_array
,
///< the character for array end `]`
...
@@ -9596,7 +9598,9 @@ class basic_json
...
@@ -9596,7 +9598,9 @@ class basic_json
return
"null literal"
;
return
"null literal"
;
case
token_type
:
:
value_string
:
case
token_type
:
:
value_string
:
return
"string literal"
;
return
"string literal"
;
case
token_type
:
:
value_number
:
case
lexer
:
:
token_type
::
value_unsigned_integer
:
case
lexer
:
:
token_type
::
value_signed_integer
:
case
lexer
:
:
token_type
::
value_float
:
return
"number literal"
;
return
"number literal"
;
case
token_type
:
:
begin_array
:
case
token_type
:
:
begin_array
:
return
"'['"
;
return
"'['"
;
...
@@ -9869,11 +9873,11 @@ basic_json_parser_12:
...
@@ -9869,11 +9873,11 @@ basic_json_parser_12:
}
}
if
(
yych
<=
'0'
)
if
(
yych
<=
'0'
)
{
{
goto
basic_json_parser_
1
3
;
goto
basic_json_parser_
4
3
;
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_
1
5
;
goto
basic_json_parser_
4
5
;
}
}
goto
basic_json_parser_5
;
goto
basic_json_parser_5
;
basic_json_parser_13:
basic_json_parser_13:
...
@@ -9883,23 +9887,23 @@ basic_json_parser_13:
...
@@ -9883,23 +9887,23 @@ basic_json_parser_13:
{
{
if
(
yych
==
'.'
)
if
(
yych
==
'.'
)
{
{
goto
basic_json_parser_4
3
;
goto
basic_json_parser_4
7
;
}
}
}
}
else
else
{
{
if
(
yych
<=
'E'
)
if
(
yych
<=
'E'
)
{
{
goto
basic_json_parser_4
4
;
goto
basic_json_parser_4
8
;
}
}
if
(
yych
==
'e'
)
if
(
yych
==
'e'
)
{
{
goto
basic_json_parser_4
4
;
goto
basic_json_parser_4
8
;
}
}
}
}
basic_json_parser_14:
basic_json_parser_14:
{
{
last_token_type
=
token_type
::
value_
numb
er
;
last_token_type
=
token_type
::
value_
unsigned_integ
er
;
break
;
break
;
}
}
basic_json_parser_15:
basic_json_parser_15:
...
@@ -9918,7 +9922,7 @@ basic_json_parser_15:
...
@@ -9918,7 +9922,7 @@ basic_json_parser_15:
{
{
if
(
yych
==
'.'
)
if
(
yych
==
'.'
)
{
{
goto
basic_json_parser_4
3
;
goto
basic_json_parser_4
7
;
}
}
goto
basic_json_parser_14
;
goto
basic_json_parser_14
;
}
}
...
@@ -9926,11 +9930,11 @@ basic_json_parser_15:
...
@@ -9926,11 +9930,11 @@ basic_json_parser_15:
{
{
if
(
yych
<=
'E'
)
if
(
yych
<=
'E'
)
{
{
goto
basic_json_parser_4
4
;
goto
basic_json_parser_4
8
;
}
}
if
(
yych
==
'e'
)
if
(
yych
==
'e'
)
{
{
goto
basic_json_parser_4
4
;
goto
basic_json_parser_4
8
;
}
}
goto
basic_json_parser_14
;
goto
basic_json_parser_14
;
}
}
...
@@ -9957,7 +9961,7 @@ basic_json_parser_23:
...
@@ -9957,7 +9961,7 @@ basic_json_parser_23:
yych
=
*
(
m_marker
=
++
m_cursor
);
yych
=
*
(
m_marker
=
++
m_cursor
);
if
(
yych
==
'a'
)
if
(
yych
==
'a'
)
{
{
goto
basic_json_parser_4
5
;
goto
basic_json_parser_4
9
;
}
}
goto
basic_json_parser_5
;
goto
basic_json_parser_5
;
basic_json_parser_24:
basic_json_parser_24:
...
@@ -9965,7 +9969,7 @@ basic_json_parser_24:
...
@@ -9965,7 +9969,7 @@ basic_json_parser_24:
yych
=
*
(
m_marker
=
++
m_cursor
);
yych
=
*
(
m_marker
=
++
m_cursor
);
if
(
yych
==
'u'
)
if
(
yych
==
'u'
)
{
{
goto
basic_json_parser_
46
;
goto
basic_json_parser_
50
;
}
}
goto
basic_json_parser_5
;
goto
basic_json_parser_5
;
basic_json_parser_25:
basic_json_parser_25:
...
@@ -9973,7 +9977,7 @@ basic_json_parser_25:
...
@@ -9973,7 +9977,7 @@ basic_json_parser_25:
yych
=
*
(
m_marker
=
++
m_cursor
);
yych
=
*
(
m_marker
=
++
m_cursor
);
if
(
yych
==
'r'
)
if
(
yych
==
'r'
)
{
{
goto
basic_json_parser_
47
;
goto
basic_json_parser_
51
;
}
}
goto
basic_json_parser_5
;
goto
basic_json_parser_5
;
basic_json_parser_26:
basic_json_parser_26:
...
@@ -10055,6 +10059,8 @@ basic_json_parser_31:
...
@@ -10055,6 +10059,8 @@ basic_json_parser_31:
}
}
basic_json_parser_32:
basic_json_parser_32:
m_cursor
=
m_marker
;
m_cursor
=
m_marker
;
if
(
yyaccept
<=
1
)
{
if
(
yyaccept
==
0
)
if
(
yyaccept
==
0
)
{
{
goto
basic_json_parser_5
;
goto
basic_json_parser_5
;
...
@@ -10063,6 +10069,18 @@ basic_json_parser_32:
...
@@ -10063,6 +10069,18 @@ basic_json_parser_32:
{
{
goto
basic_json_parser_14
;
goto
basic_json_parser_14
;
}
}
}
else
{
if
(
yyaccept
==
2
)
{
goto
basic_json_parser_44
;
}
else
{
goto
basic_json_parser_55
;
}
}
basic_json_parser_33:
basic_json_parser_33:
++
m_cursor
;
++
m_cursor
;
{
{
...
@@ -10142,7 +10160,7 @@ basic_json_parser_35:
...
@@ -10142,7 +10160,7 @@ basic_json_parser_35:
}
}
if
(
yych
<=
'u'
)
if
(
yych
<=
'u'
)
{
{
goto
basic_json_parser_
48
;
goto
basic_json_parser_
52
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
...
@@ -10261,6 +10279,71 @@ basic_json_parser_42:
...
@@ -10261,6 +10279,71 @@ basic_json_parser_42:
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_43:
basic_json_parser_43:
yyaccept
=
2
;
yych
=
*
(
m_marker
=
++
m_cursor
);
if
(
yych
<=
'D'
)
{
if
(
yych
==
'.'
)
{
goto
basic_json_parser_47
;
}
}
else
{
if
(
yych
<=
'E'
)
{
goto
basic_json_parser_48
;
}
if
(
yych
==
'e'
)
{
goto
basic_json_parser_48
;
}
}
basic_json_parser_44:
{
last_token_type
=
token_type
::
value_signed_integer
;
break
;
}
basic_json_parser_45:
yyaccept
=
2
;
m_marker
=
++
m_cursor
;
if
((
m_limit
-
m_cursor
)
<
3
)
{
fill_line_buffer
(
3
);
// LCOV_EXCL_LINE
}
yych
=
*
m_cursor
;
if
(
yych
<=
'9'
)
{
if
(
yych
==
'.'
)
{
goto
basic_json_parser_47
;
}
if
(
yych
<=
'/'
)
{
goto
basic_json_parser_44
;
}
goto
basic_json_parser_45
;
}
else
{
if
(
yych
<=
'E'
)
{
if
(
yych
<=
'D'
)
{
goto
basic_json_parser_44
;
}
goto
basic_json_parser_48
;
}
else
{
if
(
yych
==
'e'
)
{
goto
basic_json_parser_48
;
}
goto
basic_json_parser_44
;
}
}
basic_json_parser_47:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
<=
'/'
)
if
(
yych
<=
'/'
)
{
{
...
@@ -10268,16 +10351,16 @@ basic_json_parser_43:
...
@@ -10268,16 +10351,16 @@ basic_json_parser_43:
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_
49
;
goto
basic_json_parser_
53
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_4
4
:
basic_json_parser_4
8
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
<=
','
)
if
(
yych
<=
','
)
{
{
if
(
yych
==
'+'
)
if
(
yych
==
'+'
)
{
{
goto
basic_json_parser_5
1
;
goto
basic_json_parser_5
6
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
...
@@ -10285,7 +10368,7 @@ basic_json_parser_44:
...
@@ -10285,7 +10368,7 @@ basic_json_parser_44:
{
{
if
(
yych
<=
'-'
)
if
(
yych
<=
'-'
)
{
{
goto
basic_json_parser_5
1
;
goto
basic_json_parser_5
6
;
}
}
if
(
yych
<=
'/'
)
if
(
yych
<=
'/'
)
{
{
...
@@ -10293,32 +10376,32 @@ basic_json_parser_44:
...
@@ -10293,32 +10376,32 @@ basic_json_parser_44:
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_5
2
;
goto
basic_json_parser_5
7
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
basic_json_parser_4
5
:
basic_json_parser_4
9
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
==
'l'
)
if
(
yych
==
'l'
)
{
{
goto
basic_json_parser_5
4
;
goto
basic_json_parser_5
9
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_
46
:
basic_json_parser_
50
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
==
'l'
)
if
(
yych
==
'l'
)
{
{
goto
basic_json_parser_
55
;
goto
basic_json_parser_
60
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_
47
:
basic_json_parser_
51
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
==
'u'
)
if
(
yych
==
'u'
)
{
{
goto
basic_json_parser_
56
;
goto
basic_json_parser_
61
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_
48
:
basic_json_parser_
52
:
++
m_cursor
;
++
m_cursor
;
if
(
m_limit
<=
m_cursor
)
if
(
m_limit
<=
m_cursor
)
{
{
...
@@ -10333,7 +10416,7 @@ basic_json_parser_48:
...
@@ -10333,7 +10416,7 @@ basic_json_parser_48:
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_
57
;
goto
basic_json_parser_
62
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
...
@@ -10341,7 +10424,7 @@ basic_json_parser_48:
...
@@ -10341,7 +10424,7 @@ basic_json_parser_48:
{
{
if
(
yych
<=
'F'
)
if
(
yych
<=
'F'
)
{
{
goto
basic_json_parser_
57
;
goto
basic_json_parser_
62
;
}
}
if
(
yych
<=
'`'
)
if
(
yych
<=
'`'
)
{
{
...
@@ -10349,12 +10432,12 @@ basic_json_parser_48:
...
@@ -10349,12 +10432,12 @@ basic_json_parser_48:
}
}
if
(
yych
<=
'f'
)
if
(
yych
<=
'f'
)
{
{
goto
basic_json_parser_
57
;
goto
basic_json_parser_
62
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
basic_json_parser_
49
:
basic_json_parser_
53
:
yyaccept
=
1
;
yyaccept
=
3
;
m_marker
=
++
m_cursor
;
m_marker
=
++
m_cursor
;
if
((
m_limit
-
m_cursor
)
<
3
)
if
((
m_limit
-
m_cursor
)
<
3
)
{
{
...
@@ -10365,27 +10448,30 @@ basic_json_parser_49:
...
@@ -10365,27 +10448,30 @@ basic_json_parser_49:
{
{
if
(
yych
<=
'/'
)
if
(
yych
<=
'/'
)
{
{
goto
basic_json_parser_
14
;
goto
basic_json_parser_
55
;
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_
49
;
goto
basic_json_parser_
53
;
}
}
goto
basic_json_parser_14
;
}
}
else
else
{
{
if
(
yych
<=
'E'
)
if
(
yych
<=
'E'
)
{
{
goto
basic_json_parser_4
4
;
goto
basic_json_parser_4
8
;
}
}
if
(
yych
==
'e'
)
if
(
yych
==
'e'
)
{
{
goto
basic_json_parser_4
4
;
goto
basic_json_parser_4
8
;
}
}
goto
basic_json_parser_14
;
}
}
basic_json_parser_51:
basic_json_parser_55:
{
last_token_type
=
token_type
::
value_float
;
break
;
}
basic_json_parser_56:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
<=
'/'
)
if
(
yych
<=
'/'
)
{
{
...
@@ -10395,7 +10481,7 @@ basic_json_parser_51:
...
@@ -10395,7 +10481,7 @@ basic_json_parser_51:
{
{
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
basic_json_parser_5
2
:
basic_json_parser_5
7
:
++
m_cursor
;
++
m_cursor
;
if
(
m_limit
<=
m_cursor
)
if
(
m_limit
<=
m_cursor
)
{
{
...
@@ -10404,35 +10490,35 @@ basic_json_parser_52:
...
@@ -10404,35 +10490,35 @@ basic_json_parser_52:
yych
=
*
m_cursor
;
yych
=
*
m_cursor
;
if
(
yych
<=
'/'
)
if
(
yych
<=
'/'
)
{
{
goto
basic_json_parser_
14
;
goto
basic_json_parser_
55
;
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_5
2
;
goto
basic_json_parser_5
7
;
}
}
goto
basic_json_parser_
14
;
goto
basic_json_parser_
55
;
basic_json_parser_5
4
:
basic_json_parser_5
9
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
==
's'
)
if
(
yych
==
's'
)
{
{
goto
basic_json_parser_
58
;
goto
basic_json_parser_
63
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_
55
:
basic_json_parser_
60
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
==
'l'
)
if
(
yych
==
'l'
)
{
{
goto
basic_json_parser_
59
;
goto
basic_json_parser_
64
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_
56
:
basic_json_parser_
61
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
==
'e'
)
if
(
yych
==
'e'
)
{
{
goto
basic_json_parser_6
1
;
goto
basic_json_parser_6
6
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_
57
:
basic_json_parser_
62
:
++
m_cursor
;
++
m_cursor
;
if
(
m_limit
<=
m_cursor
)
if
(
m_limit
<=
m_cursor
)
{
{
...
@@ -10447,7 +10533,7 @@ basic_json_parser_57:
...
@@ -10447,7 +10533,7 @@ basic_json_parser_57:
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_6
3
;
goto
basic_json_parser_6
8
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
...
@@ -10455,7 +10541,7 @@ basic_json_parser_57:
...
@@ -10455,7 +10541,7 @@ basic_json_parser_57:
{
{
if
(
yych
<=
'F'
)
if
(
yych
<=
'F'
)
{
{
goto
basic_json_parser_6
3
;
goto
basic_json_parser_6
8
;
}
}
if
(
yych
<=
'`'
)
if
(
yych
<=
'`'
)
{
{
...
@@ -10463,30 +10549,30 @@ basic_json_parser_57:
...
@@ -10463,30 +10549,30 @@ basic_json_parser_57:
}
}
if
(
yych
<=
'f'
)
if
(
yych
<=
'f'
)
{
{
goto
basic_json_parser_6
3
;
goto
basic_json_parser_6
8
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
basic_json_parser_
58
:
basic_json_parser_
63
:
yych
=
*++
m_cursor
;
yych
=
*++
m_cursor
;
if
(
yych
==
'e'
)
if
(
yych
==
'e'
)
{
{
goto
basic_json_parser_6
4
;
goto
basic_json_parser_6
9
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
basic_json_parser_
59
:
basic_json_parser_
64
:
++
m_cursor
;
++
m_cursor
;
{
{
last_token_type
=
token_type
::
literal_null
;
last_token_type
=
token_type
::
literal_null
;
break
;
break
;
}
}
basic_json_parser_6
1
:
basic_json_parser_6
6
:
++
m_cursor
;
++
m_cursor
;
{
{
last_token_type
=
token_type
::
literal_true
;
last_token_type
=
token_type
::
literal_true
;
break
;
break
;
}
}
basic_json_parser_6
3
:
basic_json_parser_6
8
:
++
m_cursor
;
++
m_cursor
;
if
(
m_limit
<=
m_cursor
)
if
(
m_limit
<=
m_cursor
)
{
{
...
@@ -10501,7 +10587,7 @@ basic_json_parser_63:
...
@@ -10501,7 +10587,7 @@ basic_json_parser_63:
}
}
if
(
yych
<=
'9'
)
if
(
yych
<=
'9'
)
{
{
goto
basic_json_parser_
66
;
goto
basic_json_parser_
71
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
...
@@ -10509,7 +10595,7 @@ basic_json_parser_63:
...
@@ -10509,7 +10595,7 @@ basic_json_parser_63:
{
{
if
(
yych
<=
'F'
)
if
(
yych
<=
'F'
)
{
{
goto
basic_json_parser_
66
;
goto
basic_json_parser_
71
;
}
}
if
(
yych
<=
'`'
)
if
(
yych
<=
'`'
)
{
{
...
@@ -10517,17 +10603,17 @@ basic_json_parser_63:
...
@@ -10517,17 +10603,17 @@ basic_json_parser_63:
}
}
if
(
yych
<=
'f'
)
if
(
yych
<=
'f'
)
{
{
goto
basic_json_parser_
66
;
goto
basic_json_parser_
71
;
}
}
goto
basic_json_parser_32
;
goto
basic_json_parser_32
;
}
}
basic_json_parser_6
4
:
basic_json_parser_6
9
:
++
m_cursor
;
++
m_cursor
;
{
{
last_token_type
=
token_type
::
literal_false
;
last_token_type
=
token_type
::
literal_false
;
break
;
break
;
}
}
basic_json_parser_
66
:
basic_json_parser_
71
:
++
m_cursor
;
++
m_cursor
;
if
(
m_limit
<=
m_cursor
)
if
(
m_limit
<=
m_cursor
)
{
{
...
@@ -10867,42 +10953,6 @@ basic_json_parser_66:
...
@@ -10867,42 +10953,6 @@ basic_json_parser_66:
return
parse
(
val
,
std
::
is_integral
<
T
>
());
return
parse
(
val
,
std
::
is_integral
<
T
>
());
}
}
/*!
This is a helper to determine whether to parse the token into
floating-point or integral type.
@note We wouldn't need it if we had separate token types for
integral and floating-point cases.
@return true iff token matches `^[+-]\d+$`
*/
bool
is_integral
()
const
{
const
char
*
p
=
m_start
;
if
(
p
==
nullptr
)
{
return
false
;
// LCOV_EXCL_LINE
}
if
((
*
p
==
'-'
)
or
(
*
p
==
'+'
))
{
++
p
;
}
if
(
p
==
m_end
)
{
return
false
;
// LCOV_EXCL_LINE
}
while
((
p
<
m_end
)
and
(
*
p
>=
'0'
)
and
(
*
p
<=
'9'
))
{
++
p
;
}
return
(
p
==
m_end
);
}
private:
private:
const
char
*
const
m_start
=
nullptr
;
const
char
*
const
m_start
=
nullptr
;
const
char
*
const
m_end
=
nullptr
;
const
char
*
const
m_end
=
nullptr
;
...
@@ -11011,7 +11061,7 @@ basic_json_parser_66:
...
@@ -11011,7 +11061,7 @@ basic_json_parser_66:
return
(
x
==
static_cast
<
decltype
(
x
)
>
(
value
))
// x fits into destination T
return
(
x
==
static_cast
<
decltype
(
x
)
>
(
value
))
// x fits into destination T
and
(
x
<
0
)
==
(
value
<
0
)
// preserved sign
and
(
x
<
0
)
==
(
value
<
0
)
// preserved sign
and
((
x
!=
0
)
or
is_integral
())
// strto[u]ll did nto fail
//
and ((x != 0) or is_integral()) // strto[u]ll did nto fail
and
(
errno
==
0
)
// strto[u]ll did not overflow
and
(
errno
==
0
)
// strto[u]ll did not overflow
and
(
m_start
<
m_end
)
// token was not empty
and
(
m_start
<
m_end
)
// token was not empty
and
(
endptr
==
m_end
);
// parsed entire token exactly
and
(
endptr
==
m_end
);
// parsed entire token exactly
...
@@ -11035,46 +11085,56 @@ basic_json_parser_66:
...
@@ -11035,46 +11085,56 @@ basic_json_parser_66:
interpreted as a number
interpreted as a number
@param[out] result @ref basic_json object to receive the number.
@param[out] result @ref basic_json object to receive the number.
@param[in] token the type of the number token
*/
*/
void
get_number
(
basic_json
&
result
)
const
void
get_number
(
basic_json
&
result
,
const
token_type
token
)
const
{
{
assert
(
m_start
!=
nullptr
);
assert
(
m_start
!=
nullptr
);
assert
(
m_start
<
m_cursor
);
assert
(
m_start
<
m_cursor
);
assert
((
token
==
token_type
::
value_unsigned_integer
)
or
(
token
==
token_type
::
value_signed_integer
)
or
(
token
==
token_type
::
value_float
));
strtonum
num
(
reinterpret_cast
<
const
char
*>
(
m_start
),
strtonum
num
(
reinterpret_cast
<
const
char
*>
(
m_start
),
reinterpret_cast
<
const
char
*>
(
m_cursor
));
reinterpret_cast
<
const
char
*>
(
m_cursor
));
const
bool
is_negative
=
(
*
m_start
==
'-'
);
switch
(
token
)
result
.
m_type
=
value_t
::
discarded
;
if
(
not
num
.
is_integral
())
{
{
// will parse as float below
case
lexer
:
:
token_type
::
value_unsigned_integer
:
}
else
if
(
is_negative
)
{
{
number_integer
_t
val
{
0
};
number_unsigned
_t
val
{
0
};
if
(
num
.
to
(
val
))
if
(
num
.
to
(
val
))
{
{
result
.
m_type
=
value_t
::
number_integer
;
result
.
m_type
=
value_t
::
number_unsigned
;
result
.
m_value
=
val
;
result
.
m_value
=
val
;
return
;
}
}
break
;
}
}
else
case
lexer
:
:
token_type
::
value_signed_integer
:
{
{
number_unsigned
_t
val
{
0
};
number_integer
_t
val
{
0
};
if
(
num
.
to
(
val
))
if
(
num
.
to
(
val
))
{
{
result
.
m_type
=
value_t
::
number_unsigned
;
result
.
m_type
=
value_t
::
number_integer
;
result
.
m_value
=
val
;
result
.
m_value
=
val
;
return
;
}
break
;
}
default:
{
break
;
}
}
}
}
number_float_t
val
{
0
};
number_float_t
val
{
0
};
if
(
result
.
m_type
!=
value_t
::
discarded
or
(
not
num
.
to
(
val
)
))
if
(
not
num
.
to
(
val
))
{
{
// already have a value from above or couldn't parse as float_t
// couldn't parse as float_t
result
.
m_type
=
value_t
::
discarded
;
return
;
return
;
}
}
...
@@ -11330,9 +11390,11 @@ basic_json_parser_66:
...
@@ -11330,9 +11390,11 @@ basic_json_parser_66:
break
;
break
;
}
}
case
lexer
:
:
token_type
::
value_number
:
case
lexer
:
:
token_type
::
value_unsigned_integer
:
case
lexer
:
:
token_type
::
value_signed_integer
:
case
lexer
:
:
token_type
::
value_float
:
{
{
m_lexer
.
get_number
(
result
);
m_lexer
.
get_number
(
result
,
last_token
);
get_token
();
get_token
();
break
;
break
;
}
}
...
...
src/json.hpp.re2c
View file @
c2d55109
...
@@ -9444,7 +9444,9 @@ class basic_json
...
@@ -9444,7 +9444,9 @@ class basic_json
literal_false, ///< the `false` literal
literal_false, ///< the `false` literal
literal_null, ///< the `null` literal
literal_null, ///< the `null` literal
value_string, ///< a string -- use get_string() for actual value
value_string, ///< a string -- use get_string() for actual value
value_number, ///< a number -- use get_number() for actual value
value_unsigned_integer, ///< an unsigned integer -- use get_number() for actual value
value_signed_integer, ///< a signed integer -- use get_number() for actual value
value_float, ///< an floating point number -- use get_number() for actual value
begin_array, ///< the character for array begin `[`
begin_array, ///< the character for array begin `[`
begin_object, ///< the character for object begin `{`
begin_object, ///< the character for object begin `{`
end_array, ///< the character for array end `]`
end_array, ///< the character for array end `]`
...
@@ -9596,7 +9598,9 @@ class basic_json
...
@@ -9596,7 +9598,9 @@ class basic_json
return "null literal";
return "null literal";
case token_type::value_string:
case token_type::value_string:
return "string literal";
return "string literal";
case token_type::value_number:
case lexer::token_type::value_unsigned_integer:
case lexer::token_type::value_signed_integer:
case lexer::token_type::value_float:
return "number literal";
return "number literal";
case token_type::begin_array:
case token_type::begin_array:
return "'['";
return "'['";
...
@@ -9694,8 +9698,12 @@ class basic_json
...
@@ -9694,8 +9698,12 @@ class basic_json
exp = e (minus | plus)? digit+;
exp = e (minus | plus)? digit+;
frac = decimal_point digit+;
frac = decimal_point digit+;
int = (zero | digit_1_9 digit*);
int = (zero | digit_1_9 digit*);
number = minus? int frac? exp?;
number_unsigned = int;
number { last_token_type = token_type::value_number; break; }
number_unsigned { last_token_type = token_type::value_unsigned_integer; break; }
number_signed = minus int;
number_signed { last_token_type = token_type::value_signed_integer; break; }
number_float = minus? int frac? exp?;
number_float { last_token_type = token_type::value_float; break; }
// string
// string
quotation_mark = "\"";
quotation_mark = "\"";
...
@@ -10017,42 +10025,6 @@ class basic_json
...
@@ -10017,42 +10025,6 @@ class basic_json
return parse(val, std::is_integral<T>());
return parse(val, std::is_integral<T>());
}
}
/*!
This is a helper to determine whether to parse the token into
floating-point or integral type.
@note We wouldn't need it if we had separate token types for
integral and floating-point cases.
@return true iff token matches `^[+-]\d+$`
*/
bool is_integral() const
{
const char* p = m_start;
if (p == nullptr)
{
return false; // LCOV_EXCL_LINE
}
if ((*p == '-') or (*p == '+'))
{
++p;
}
if (p == m_end)
{
return false; // LCOV_EXCL_LINE
}
while ((p < m_end) and (*p >= '0') and (*p <= '9'))
{
++p;
}
return (p == m_end);
}
private:
private:
const char* const m_start = nullptr;
const char* const m_start = nullptr;
const char* const m_end = nullptr;
const char* const m_end = nullptr;
...
@@ -10161,7 +10133,7 @@ class basic_json
...
@@ -10161,7 +10133,7 @@ class basic_json
return (x == static_cast<decltype(x)>(value)) // x fits into destination T
return (x == static_cast<decltype(x)>(value)) // x fits into destination T
and (x < 0) == (value < 0) // preserved sign
and (x < 0) == (value < 0) // preserved sign
and ((x != 0) or is_integral()) // strto[u]ll did nto fail
//
and ((x != 0) or is_integral()) // strto[u]ll did nto fail
and (errno == 0) // strto[u]ll did not overflow
and (errno == 0) // strto[u]ll did not overflow
and (m_start < m_end) // token was not empty
and (m_start < m_end) // token was not empty
and (endptr == m_end); // parsed entire token exactly
and (endptr == m_end); // parsed entire token exactly
...
@@ -10185,46 +10157,56 @@ class basic_json
...
@@ -10185,46 +10157,56 @@ class basic_json
interpreted as a number
interpreted as a number
@param[out] result @ref basic_json object to receive the number.
@param[out] result @ref basic_json object to receive the number.
@param[in] token the type of the number token
*/
*/
void get_number(basic_json& result) const
void get_number(basic_json& result
, const token_type token
) const
{
{
assert(m_start != nullptr);
assert(m_start != nullptr);
assert(m_start < m_cursor);
assert(m_start < m_cursor);
assert((token == token_type::value_unsigned_integer) or
(token == token_type::value_signed_integer) or
(token == token_type::value_float));
strtonum num(reinterpret_cast<const char*>(m_start),
strtonum num(reinterpret_cast<const char*>(m_start),
reinterpret_cast<const char*>(m_cursor));
reinterpret_cast<const char*>(m_cursor));
const bool is_negative = (*m_start == '-');
switch (token)
result.m_type = value_t::discarded;
if (not num.is_integral())
{
{
// will parse as float below
case lexer::token_type::value_unsigned_integer:
}
else if (is_negative)
{
{
number_integer
_t val{0};
number_unsigned
_t val{0};
if (num.to(val))
if (num.to(val))
{
{
result.m_type = value_t::number_integer
;
result.m_type = value_t::number_unsigned
;
result.m_value = val;
result.m_value = val;
return;
}
}
break;
}
}
else
case lexer::token_type::value_signed_integer:
{
{
number_unsigned
_t val{0};
number_integer
_t val{0};
if (num.to(val))
if (num.to(val))
{
{
result.m_type = value_t::number_unsigned
;
result.m_type = value_t::number_integer
;
result.m_value = val;
result.m_value = val;
return;
}
break;
}
default:
{
break;
}
}
}
}
number_float_t val{0};
number_float_t val{0};
if (
result.m_type != value_t::discarded or (not num.to(val)
))
if (
not num.to(val
))
{
{
// already have a value from above or couldn't parse as float_t
// couldn't parse as float_t
result.m_type = value_t::discarded;
return;
return;
}
}
...
@@ -10480,9 +10462,11 @@ class basic_json
...
@@ -10480,9 +10462,11 @@ class basic_json
break;
break;
}
}
case lexer::token_type::value_number:
case lexer::token_type::value_unsigned_integer:
case lexer::token_type::value_signed_integer:
case lexer::token_type::value_float:
{
{
m_lexer.get_number(result);
m_lexer.get_number(result
, last_token
);
get_token();
get_token();
break;
break;
}
}
...
...
test/src/unit-class_lexer.cpp
View file @
c2d55109
...
@@ -65,25 +65,37 @@ TEST_CASE("lexer class")
...
@@ -65,25 +65,37 @@ TEST_CASE("lexer class")
SECTION
(
"numbers"
)
SECTION
(
"numbers"
)
{
{
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"0"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"0"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"1"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"1"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"2"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"2"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"3"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"3"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"4"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"4"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"5"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"5"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"6"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"6"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"7"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"7"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"8"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"8"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
numb
er
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_
unsigned_integ
er
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"9"
),
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"9"
),
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_number
));
1
).
scan
()
==
json
::
lexer
::
token_type
::
value_unsigned_integer
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"-0"
),
2
).
scan
()
==
json
::
lexer
::
token_type
::
value_signed_integer
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"-1"
),
2
).
scan
()
==
json
::
lexer
::
token_type
::
value_signed_integer
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"1.1"
),
3
).
scan
()
==
json
::
lexer
::
token_type
::
value_float
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"-1.1"
),
4
).
scan
()
==
json
::
lexer
::
token_type
::
value_float
));
CHECK
((
json
::
lexer
(
reinterpret_cast
<
const
json
::
lexer
::
lexer_char_t
*>
(
"1E10"
),
4
).
scan
()
==
json
::
lexer
::
token_type
::
value_float
));
}
}
SECTION
(
"whitespace"
)
SECTION
(
"whitespace"
)
...
@@ -109,7 +121,9 @@ TEST_CASE("lexer class")
...
@@ -109,7 +121,9 @@ TEST_CASE("lexer class")
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
literal_false
)
==
"false literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
literal_false
)
==
"false literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
literal_null
)
==
"null literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
literal_null
)
==
"null literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
value_string
)
==
"string literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
value_string
)
==
"string literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
value_number
)
==
"number literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
value_unsigned_integer
)
==
"number literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
value_signed_integer
)
==
"number literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
value_float
)
==
"number literal"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
begin_array
)
==
"'['"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
begin_array
)
==
"'['"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
begin_object
)
==
"'{'"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
begin_object
)
==
"'{'"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
end_array
)
==
"']'"
));
CHECK
((
json
::
lexer
::
token_type_name
(
json
::
lexer
::
token_type
::
end_array
)
==
"']'"
));
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment