Deleted preprocesser

Will be reworking it later
This commit is contained in:
2024-06-01 14:37:16 +01:00
parent 683636c66d
commit 42dbf515f0
4 changed files with 2 additions and 316 deletions

View File

@@ -15,7 +15,7 @@ TERM_RESET:=$(shell echo -e "\e[0;0m")
# Setup variables for source code, output, etc
## ASSEMBLY setup
SRC=src
CODE:=$(addprefix $(SRC)/, base.cpp lexer.cpp preprocesser.cpp)
CODE:=$(addprefix $(SRC)/, base.cpp lexer.cpp)
OBJECTS:=$(CODE:$(SRC)/%.cpp=$(DIST)/%.o)
OUT=$(DIST)/asm.out

View File

@@ -10,12 +10,10 @@
* Description: Entrypoint for assembly program
*/
#include <algorithm>
#include <cstdio>
#include <iostream>
#include <optional>
#include <string>
#include <tuple>
#include <vector>
extern "C"
@@ -25,10 +23,9 @@ extern "C"
#include "./base.hpp"
#include "./lexer.hpp"
#include "./preprocesser.hpp"
using std::cout, std::cerr, std::endl;
using std::pair, std::string, std::string_view, std::vector;
using std::string, std::string_view, std::vector;
using Lexer::Token;
using Lex_Err = Lexer::Err;
@@ -71,7 +68,6 @@ int main(int argc, const char *argv[])
string_view src;
vector<Token *> tokens, preprocessed_tokens;
Lex_Err lerr;
pp_err_t pp_err;
// Highest scoped variable cut off point
@@ -113,39 +109,9 @@ int main(int argc, const char *argv[])
#endif
}
// preprocessing
pp_err = preprocesser(tokens, preprocessed_tokens);
if (pp_err.type != pp_err_type_t::OK)
{
cerr << source_name << ":" << pp_err.reference->line << ":"
<< pp_err.reference->column << ": " << pp_err << endl;
ret = 255 - static_cast<int>(pp_err.type);
goto end;
}
else
{
#if VERBOSE >= 1
printf("[%sPREPROCESSOR%s]: %lu tokens -> %lu tokens\n", TERM_GREEN,
TERM_RESET, tokens.size(), preprocessed_tokens.size());
#endif
#if VERBOSE == 2
printf("[%sPREPROCESSOR%s]: Processed tokens: "
"\n-----------------------------------------------------------------"
"---------------\n",
TERM_GREEN, TERM_RESET);
for (auto token : preprocessed_tokens)
cout << "\t" << *token << endl;
printf("-------------------------------------------------------------"
"-------------------\n");
#endif
}
end:
for (auto token : tokens)
delete token;
for (auto token : preprocessed_tokens)
delete token;
return ret;
}

View File

@@ -1,217 +0,0 @@
/* Copyright (C) 2024 Aryadev Chavali
* You may distribute and modify this code under the terms of the
* GPLv2 license. You should have received a copy of the GPLv2
* license with this file. If not, please write to:
* aryadev@aryadevchavali.com.
* Created: 2024-04-14
* Author: Aryadev Chavali
* Description: Preprocessor which occurs after lexing before parsing.
*/
#include "./preprocesser.hpp"
#include "./base.hpp"
#include <algorithm>
#include <unordered_map>
using std::pair, std::vector, std::make_pair, std::string, std::string_view;
using Lexer::Token, Lexer::Err;
#define VCLEAR(V) \
std::for_each((V).begin(), (V).end(), \
[](Token *t) \
{ \
delete t; \
});
pp_err_t preprocess_use_blocks(const vector<Token *> &tokens,
vector<Token *> &vec_out)
{
for (size_t i = 0; i < tokens.size(); ++i)
{
Token *t = tokens[i];
if (t->type == Token::Type::PP_USE)
{
if (i + 1 >= tokens.size() ||
tokens[i + 1]->type != Token::Type::LITERAL_STRING)
{
VCLEAR(vec_out);
vec_out.clear();
return pp_err_t(pp_err_type_t::EXPECTED_STRING, t);
}
Token *name = tokens[i + 1];
auto source = read_file(name->content.c_str());
if (!source)
{
VCLEAR(vec_out);
vec_out.clear();
return pp_err_t(pp_err_type_t::FILE_NONEXISTENT, name);
}
std::vector<Token *> ftokens;
Err lerr = tokenise_buffer(source.value(), ftokens);
if (lerr.type != Err::Type::OK)
{
VCLEAR(vec_out);
vec_out.clear();
return pp_err_t(pp_err_type_t::FILE_PARSE_ERROR, name, lerr);
}
vec_out.insert(vec_out.end(), ftokens.begin(), ftokens.end());
++i;
}
else
vec_out.push_back(new Token{*t});
}
return pp_err_t();
}
struct const_t
{
size_t start, end;
};
pp_err_t preprocess_const_blocks(const vector<Token *> &tokens,
vector<Token *> &vec_out)
{
std::unordered_map<string_view, const_t> blocks;
for (size_t i = 0; i < tokens.size(); ++i)
{
Token *t = tokens[i];
if (t->type == Token::Type::PP_CONST)
{
string_view capture;
if (i + 1 >= tokens.size() || tokens[i + 1]->type != Token::Type::SYMBOL)
return pp_err_type_t::EXPECTED_NAME;
capture = tokens[++i]->content;
++i;
size_t block_start = i, block_end = 0;
for (; i < tokens.size() && tokens[i]->type != Token::Type::PP_END; ++i)
continue;
if (i == tokens.size())
return pp_err_t{pp_err_type_t::EXPECTED_END};
block_end = i;
blocks[capture] = const_t{block_start, block_end};
}
}
if (blocks.size() == 0)
{
// Just construct a new vector and carry on
for (Token *token : tokens)
vec_out.push_back(new Token{*token});
}
else
{
for (size_t i = 0; i < tokens.size(); ++i)
{
Token *token = tokens[i];
// Skip the tokens that construct the const
if (token->type == Token::Type::PP_CONST)
for (; i < tokens.size() && tokens[i]->type != Token::Type::PP_END; ++i)
continue;
else if (token->type == Token::Type::PP_REFERENCE)
{
auto it = blocks.find(token->content);
if (it == blocks.end())
{
VCLEAR(vec_out);
vec_out.clear();
return pp_err_t(pp_err_type_t::UNKNOWN_NAME, token);
}
const_t block = it->second;
for (size_t i = block.start; i < block.end; ++i)
vec_out.push_back(new Token{*tokens[i]});
}
else
vec_out.push_back(new Token{*token});
}
}
return pp_err_t();
}
pp_err_t preprocesser(const vector<Token *> &tokens, vector<Token *> &vec_out)
{
vector<Token *> use_block_tokens;
pp_err_t pperr = preprocess_use_blocks(tokens, use_block_tokens);
if (pperr.type != pp_err_type_t::OK)
{
vec_out = tokens;
return pperr;
}
vector<Token *> const_block_tokens;
pperr = preprocess_const_blocks(use_block_tokens, const_block_tokens);
if (pperr.type != pp_err_type_t::OK)
{
VCLEAR(tokens);
vec_out = use_block_tokens;
return pperr;
}
VCLEAR(use_block_tokens);
vec_out = const_block_tokens;
return pp_err_t{pp_err_type_t::OK};
}
// TODO: Implement this
pp_err_t preprocess_macro_blocks(const vector<Token *> &, vector<Token *> &);
std::ostream &operator<<(std::ostream &os, pp_err_t &err)
{
os << "PREPROCESSING_";
switch (err.type)
{
case OK:
return os << "OK";
case EXPECTED_NAME:
return os << "EXPECTED_NAME";
case EXPECTED_STRING:
return os << "EXPECTED_STRING";
case EXPECTED_END:
return os << "EXPECTED_END";
case FILE_NONEXISTENT:
return os << "FILE_NONEXISTENT";
case FILE_PARSE_ERROR:
return os << "FILE_PARSE_ERROR -> \n\t[" << err.reference->content
<< "]:" << err.lerr;
case UNKNOWN_NAME:
return os << "UNKNOWN_NAME";
}
return os;
}
pp_err_t::pp_err_t() : reference{nullptr}, type{pp_err_type_t::OK}, lerr{}
{}
pp_err_t::pp_err_t(pp_err_type_t e) : reference{nullptr}, type{e}, lerr{}
{}
pp_err_t::pp_err_t(pp_err_type_t err, const Token *ref)
: reference{ref}, type{err}
{}
pp_err_t::pp_err_t(pp_err_type_t err, const Token *ref, Err lerr)
: reference{ref}, type{err}, lerr{lerr}
{}
// pp_unit_t::pp_unit_t(const Token *const token) : resolved{false},
// token{token}
// {}
// pp_unit_t::pp_unit_t(std::string_view name, std::vector<pp_unit_t> elements)
// : resolved{false}, token{nullptr}, container{name, elements}
// {}

View File

@@ -1,63 +0,0 @@
/* Copyright (C) 2024 Aryadev Chavali
* You may distribute and modify this code under the terms of the GPLv2
* license. You should have received a copy of the GPLv2 license with
* this file. If not, please write to: aryadev@aryadevchavali.com.
* Created: 2024-04-14
* Author: Aryadev Chavali
* Description: Preprocessor which occurs after lexing before parsing.
*/
#ifndef PREPROCESSER_HPP
#define PREPROCESSER_HPP
#include <ostream>
#include <tuple>
#include "./lexer.hpp"
enum pp_err_type_t
{
OK = 0,
EXPECTED_NAME,
EXPECTED_STRING,
EXPECTED_END,
FILE_NONEXISTENT,
FILE_PARSE_ERROR,
UNKNOWN_NAME,
};
struct pp_err_t
{
const Lexer::Token *reference;
pp_err_type_t type;
Lexer::Err lerr;
pp_err_t();
pp_err_t(pp_err_type_t);
pp_err_t(pp_err_type_t, const Lexer::Token *);
pp_err_t(pp_err_type_t, const Lexer::Token *, Lexer::Err);
};
std::ostream &operator<<(std::ostream &, pp_err_t &);
struct pp_unit_t
{
const Lexer::Token *const token;
struct
{
std::string_view name;
std::vector<pp_unit_t> elements;
} container;
pp_unit_t(const Lexer::Token *const);
pp_unit_t(std::string_view, std::vector<pp_unit_t>);
};
std::vector<pp_unit_t> tokens_to_units(const std::vector<Lexer::Token *> &);
pp_err_t preprocess_use(std::vector<pp_unit_t> &);
pp_err_t preprocesser(const std::vector<Lexer::Token *> &,
std::vector<Lexer::Token *> &);
#endif