矢量器标准化中无效的模式
我正在尝试将独特的字符与单词分开,以便我可以将它们视为单独的令牌。
这是我的矢量器:
vectorizer = tf.keras.layers.TextVectorization(
standardize=code_standaridization,
split="whitespace",
output_mode="int",
output_sequence_length=params["input_dim"],
)
这是我的功能:
def code_standaridization(input_data):
"""
Helps with finding correct embedding.
"""
input_data = tf.strings.regex_replace(input_data, "-", " - ")
input_data = tf.strings.regex_replace(input_data, "_", " _ ")
input_data = tf.strings.regex_replace(input_data, "(", " ( ")
input_data = tf.strings.regex_replace(input_data, ")", " ) ")
input_data = tf.strings.regex_replace(input_data, "{", " { ")
input_data = tf.strings.regex_replace(input_data, "}", " } ")
input_data = tf.strings.regex_replace(input_data, "[", " [ ")
input_data = tf.strings.regex_replace(input_data, "]", " ] ")
input_data = tf.strings.regex_replace(input_data, '"', ' " ')
input_data = tf.strings.regex_replace(input_data, "'", " ' ")
input_data = tf.strings.regex_replace(input_data, ".", " . ")
input_data = tf.strings.regex_replace(input_data, ",", " , ")
return input_data
但是我遇到了这个错误:
Node: 'StaticRegexReplace_2'
Invalid pattern: (, error: missing ): (
[[{{node StaticRegexReplace_2}}]] [Op:__inference_adapt_step_133]
I am trying to separate unique characters from words so I can threat them as separate tokens.
This is my vectorizer:
vectorizer = tf.keras.layers.TextVectorization(
standardize=code_standaridization,
split="whitespace",
output_mode="int",
output_sequence_length=params["input_dim"],
)
And this is my function:
def code_standaridization(input_data):
"""
Helps with finding correct embedding.
"""
input_data = tf.strings.regex_replace(input_data, "-", " - ")
input_data = tf.strings.regex_replace(input_data, "_", " _ ")
input_data = tf.strings.regex_replace(input_data, "(", " ( ")
input_data = tf.strings.regex_replace(input_data, ")", " ) ")
input_data = tf.strings.regex_replace(input_data, "{", " { ")
input_data = tf.strings.regex_replace(input_data, "}", " } ")
input_data = tf.strings.regex_replace(input_data, "[", " [ ")
input_data = tf.strings.regex_replace(input_data, "]", " ] ")
input_data = tf.strings.regex_replace(input_data, '"', ' " ')
input_data = tf.strings.regex_replace(input_data, "'", " ' ")
input_data = tf.strings.regex_replace(input_data, ".", " . ")
input_data = tf.strings.regex_replace(input_data, ",", " , ")
return input_data
But I am getting this error:
Node: 'StaticRegexReplace_2'
Invalid pattern: (, error: missing ): (
[[{{node StaticRegexReplace_2}}]] [Op:__inference_adapt_step_133]
如果你对这篇内容有疑问,欢迎到本站社区发帖提问 参与讨论,获取更多帮助,或者扫码二维码加入 Web 技术交流群。

绑定邮箱获取回复消息
由于您还没有绑定你的真实邮箱,如果其他用户或者作者回复了您的评论,将不能在第一时间通知您!
发布评论
评论(1)
regex_replace(Regex_replace)
定义在这里,哪个注释:因此,您应该添加后斜切以更换括号:
regex_replace()
uses the RE2 syntax defined here, which notes:So, you should add backslashes for replacing parentheses: